ansible-playbook 2.9.27
config file = None
configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/local/lib/python3.9/site-packages/ansible
executable location = /usr/local/bin/ansible-playbook
python version = 3.9.19 (main, May 16 2024, 11:40:09) [GCC 8.5.0 20210514 (Red Hat 8.5.0-22)]
No config file found; using defaults
[WARNING]: running playbook inside collection fedora.linux_system_roles
Skipping callback 'actionable', as we already have a stdout callback.
Skipping callback 'counter_enabled', as we already have a stdout callback.
Skipping callback 'debug', as we already have a stdout callback.
Skipping callback 'dense', as we already have a stdout callback.
Skipping callback 'dense', as we already have a stdout callback.
Skipping callback 'full_skip', as we already have a stdout callback.
Skipping callback 'json', as we already have a stdout callback.
Skipping callback 'minimal', as we already have a stdout callback.
Skipping callback 'null', as we already have a stdout callback.
Skipping callback 'oneline', as we already have a stdout callback.
Skipping callback 'selective', as we already have a stdout callback.
Skipping callback 'skippy', as we already have a stdout callback.
Skipping callback 'stderr', as we already have a stdout callback.
Skipping callback 'unixy', as we already have a stdout callback.
Skipping callback 'yaml', as we already have a stdout callback.
PLAYBOOK: tests_sbd_defaults.yml ***********************************************
2 plays in /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml
PLAY [all] *********************************************************************
META: ran handlers
TASK [Include vault variables] *************************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:5
Monday 29 July 2024 08:17:54 -0400 (0:00:00.017) 0:00:00.017 ***********
ok: [managed_node1] => {
"ansible_facts": {
"ha_cluster_hacluster_password": {
"__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n31303833633366333561656439323930303361333161363239346166656537323933313436\n3432386236656563343237306335323637396239616230353561330a313731623238393238\n62343064666336643930663239383936616465643134646536656532323461356237646133\n3761616633323839633232353637366266350a313163633236376666653238633435306565\n3264623032333736393535663833\n"
}
},
"ansible_included_var_files": [
"/var/ARTIFACTS/work-generalwpbd9hvt/plans/general/tree/tests/vars/vault-variables.yml"
],
"changed": false
}
META: ran handlers
META: ran handlers
PLAY [Minimal SBD configuration] ***********************************************
TASK [Gathering Facts] *********************************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:9
Monday 29 July 2024 08:17:54 -0400 (0:00:00.016) 0:00:00.033 ***********
ok: [managed_node1]
META: ran handlers
TASK [Set up test environment] *************************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:20
Monday 29 July 2024 08:17:55 -0400 (0:00:00.960) 0:00:00.994 ***********
TASK [fedora.linux_system_roles.ha_cluster : Set node name to 'localhost' for single-node clusters] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:9
Monday 29 July 2024 08:17:55 -0400 (0:00:00.027) 0:00:01.022 ***********
ok: [managed_node1] => {
"ansible_facts": {
"inventory_hostname": "localhost"
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Ensure facts used by tests] *******
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:14
Monday 29 July 2024 08:17:55 -0400 (0:00:00.029) 0:00:01.051 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Check if system is ostree] ********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:22
Monday 29 July 2024 08:17:55 -0400 (0:00:00.027) 0:00:01.078 ***********
ok: [managed_node1] => {
"changed": false,
"stat": {
"exists": false
}
}
TASK [fedora.linux_system_roles.ha_cluster : Set flag to indicate system is ostree] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:27
Monday 29 July 2024 08:17:55 -0400 (0:00:00.444) 0:00:01.523 ***********
ok: [managed_node1] => {
"ansible_facts": {
"__ha_cluster_is_ostree": false
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Do not try to enable RHEL repositories] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:32
Monday 29 July 2024 08:17:55 -0400 (0:00:00.031) 0:00:01.554 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Copy nss-altfiles ha_cluster users to /etc/passwd] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:41
Monday 29 July 2024 08:17:56 -0400 (0:00:00.056) 0:00:01.611 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Set up test environment for SBD] *****************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:25
Monday 29 July 2024 08:17:56 -0400 (0:00:00.026) 0:00:01.638 ***********
TASK [fedora.linux_system_roles.ha_cluster : Load softdog module for SBD to have at least one watchdog] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup_sbd.yml:9
Monday 29 July 2024 08:17:56 -0400 (0:00:00.025) 0:00:01.663 ***********
changed: [managed_node1] => {
"changed": true,
"cmd": [
"modprobe",
"softdog"
],
"delta": "0:00:00.047694",
"end": "2024-07-29 08:17:56.497320",
"rc": 0,
"start": "2024-07-29 08:17:56.449626"
}
TASK [fedora.linux_system_roles.ha_cluster : Create backing files for SBD devices] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup_sbd.yml:15
Monday 29 July 2024 08:17:56 -0400 (0:00:00.475) 0:00:02.139 ***********
changed: [managed_node1] => {
"changed": true,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.qz3pd4ht_ha_cluster_tests",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Initialize backing files for SBD devices] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup_sbd.yml:21
Monday 29 July 2024 08:17:56 -0400 (0:00:00.406) 0:00:02.545 ***********
changed: [managed_node1] => {
"changed": true,
"cmd": [
"dd",
"if=/dev/zero",
"of=/tmp/ansible.qz3pd4ht_ha_cluster_tests",
"bs=1M",
"count=10"
],
"delta": "0:00:00.011225",
"end": "2024-07-29 08:17:57.273128",
"rc": 0,
"start": "2024-07-29 08:17:57.261903"
}
STDERR:
10+0 records in
10+0 records out
10485760 bytes (10 MB, 10 MiB) copied, 0.00421433 s, 2.5 GB/s
TASK [fedora.linux_system_roles.ha_cluster : Mount SBD devices] ****************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup_sbd.yml:26
Monday 29 July 2024 08:17:57 -0400 (0:00:00.364) 0:00:02.910 ***********
changed: [managed_node1] => {
"changed": true,
"cmd": [
"losetup",
"--show",
"--find",
"/tmp/ansible.qz3pd4ht_ha_cluster_tests"
],
"delta": "0:00:00.021747",
"end": "2024-07-29 08:17:57.645322",
"rc": 0,
"start": "2024-07-29 08:17:57.623575"
}
STDOUT:
/dev/loop0
TASK [Ensure SBD config file is not present] ***********************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:30
Monday 29 July 2024 08:17:57 -0400 (0:00:00.383) 0:00:03.293 ***********
changed: [managed_node1] => {
"changed": true,
"path": "/etc/sysconfig/sbd",
"state": "absent"
}
TASK [Run HA Cluster role] *****************************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:35
Monday 29 July 2024 08:17:58 -0400 (0:00:00.453) 0:00:03.747 ***********
TASK [fedora.linux_system_roles.ha_cluster : Set platform/version specific variables] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:3
Monday 29 July 2024 08:17:58 -0400 (0:00:00.042) 0:00:03.789 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Ensure ansible_facts used by role] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:2
Monday 29 July 2024 08:17:58 -0400 (0:00:00.021) 0:00:03.811 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Check if system is ostree] ********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:10
Monday 29 July 2024 08:17:58 -0400 (0:00:00.029) 0:00:03.840 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Set flag to indicate system is ostree] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:15
Monday 29 July 2024 08:17:58 -0400 (0:00:00.026) 0:00:03.866 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Set platform/version specific variables] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:19
Monday 29 July 2024 08:17:58 -0400 (0:00:00.027) 0:00:03.894 ***********
ok: [managed_node1] => (item=RedHat.yml) => {
"ansible_facts": {
"__ha_cluster_cloud_agents_packages": [],
"__ha_cluster_fence_agent_packages_default": "{{ ['fence-agents-all'] + (['fence-virt'] if ansible_architecture == 'x86_64' else []) }}",
"__ha_cluster_fullstack_node_packages": [
"corosync",
"libknet1-plugins-all",
"resource-agents",
"pacemaker",
"openssl"
],
"__ha_cluster_pcs_provider": "pcs-0.10",
"__ha_cluster_qdevice_node_packages": [
"corosync-qdevice",
"bash",
"coreutils",
"curl",
"grep",
"nss-tools",
"openssl",
"sed"
],
"__ha_cluster_repos": [],
"__ha_cluster_role_essential_packages": [
"pcs",
"corosync-qnetd"
],
"__ha_cluster_sbd_packages": [
"sbd"
],
"__ha_cluster_services": [
"corosync",
"corosync-qdevice",
"pacemaker"
]
},
"ansible_included_var_files": [
"/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/RedHat.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "RedHat.yml"
}
skipping: [managed_node1] => (item=CentOS.yml) => {
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS.yml",
"skip_reason": "Conditional result was False"
}
ok: [managed_node1] => (item=CentOS_8.yml) => {
"ansible_facts": {
"__ha_cluster_cloud_agents_packages": [
"resource-agents-aliyun",
"resource-agents-gcp",
"fence-agents-aliyun",
"fence-agents-aws",
"fence-agents-azure-arm",
"fence-agents-gce"
],
"__ha_cluster_repos": [
{
"id": "ha",
"name": "HighAvailability"
},
{
"id": "resilientstorage",
"name": "ResilientStorage"
}
]
},
"ansible_included_var_files": [
"/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/CentOS_8.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS_8.yml"
}
ok: [managed_node1] => (item=CentOS_8.yml) => {
"ansible_facts": {
"__ha_cluster_cloud_agents_packages": [
"resource-agents-aliyun",
"resource-agents-gcp",
"fence-agents-aliyun",
"fence-agents-aws",
"fence-agents-azure-arm",
"fence-agents-gce"
],
"__ha_cluster_repos": [
{
"id": "ha",
"name": "HighAvailability"
},
{
"id": "resilientstorage",
"name": "ResilientStorage"
}
]
},
"ansible_included_var_files": [
"/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/CentOS_8.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS_8.yml"
}
TASK [fedora.linux_system_roles.ha_cluster : Set Linux Pacemaker shell specific variables] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:34
Monday 29 July 2024 08:17:58 -0400 (0:00:00.055) 0:00:03.950 ***********
ok: [managed_node1] => {
"ansible_facts": {},
"ansible_included_var_files": [
"/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/shell_pcs.yml"
],
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Enable package repositories] ******
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:6
Monday 29 July 2024 08:17:58 -0400 (0:00:00.027) 0:00:03.977 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-package-repositories.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Find platform/version specific tasks to enable repositories] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-package-repositories.yml:3
Monday 29 July 2024 08:17:58 -0400 (0:00:00.020) 0:00:03.998 ***********
ok: [managed_node1] => (item=RedHat.yml) => {
"ansible_facts": {
"__ha_cluster_enable_repo_tasks_file": "/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/RedHat.yml"
},
"ansible_loop_var": "item",
"changed": false,
"item": "RedHat.yml"
}
ok: [managed_node1] => (item=CentOS.yml) => {
"ansible_facts": {
"__ha_cluster_enable_repo_tasks_file": "/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml"
},
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS.yml"
}
skipping: [managed_node1] => (item=CentOS_8.yml) => {
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS_8.yml",
"skip_reason": "Conditional result was False"
}
skipping: [managed_node1] => (item=CentOS_8.yml) => {
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS_8.yml",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Run platform/version specific tasks to enable repositories] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-package-repositories.yml:21
Monday 29 July 2024 08:17:58 -0400 (0:00:00.052) 0:00:04.050 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : List active CentOS repositories] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml:3
Monday 29 July 2024 08:17:58 -0400 (0:00:00.037) 0:00:04.088 ***********
[WARNING]: Consider using the dnf module rather than running 'dnf'. If you
need to use command because dnf is insufficient you can add 'warn: false' to
this command task or set 'command_warnings=False' in ansible.cfg to get rid of
this message.
ok: [managed_node1] => {
"changed": false,
"cmd": [
"dnf",
"repolist"
],
"delta": "0:00:00.235613",
"end": "2024-07-29 08:17:59.037034",
"rc": 0,
"start": "2024-07-29 08:17:58.801421"
}
STDOUT:
repo id repo name
appstream CentOS Stream 8 - AppStream
baseos CentOS Stream 8 - BaseOS
beaker-client Beaker Client - RedHatEnterpriseLinux8
beaker-harness Beaker harness
beakerlib-libraries Copr repo for beakerlib-libraries owned by bgoncalv
copr:copr.devel.redhat.com:lpol:qa-tools Copr repo for qa-tools owned by lpol
extras CentOS Stream 8 - Extras
extras-common CentOS Stream 8 - Extras common packages
ha CentOS Stream 8 - HighAvailability
TASK [fedora.linux_system_roles.ha_cluster : Enable CentOS repositories] *******
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml:10
Monday 29 July 2024 08:17:59 -0400 (0:00:00.589) 0:00:04.678 ***********
skipping: [managed_node1] => (item={'id': 'ha', 'name': 'HighAvailability'}) => {
"ansible_loop_var": "item",
"changed": false,
"item": {
"id": "ha",
"name": "HighAvailability"
},
"skip_reason": "Conditional result was False"
}
skipping: [managed_node1] => (item={'id': 'resilientstorage', 'name': 'ResilientStorage'}) => {
"ansible_loop_var": "item",
"changed": false,
"item": {
"id": "resilientstorage",
"name": "ResilientStorage"
},
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Install role essential packages] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:11
Monday 29 July 2024 08:17:59 -0400 (0:00:00.038) 0:00:04.716 ***********
ok: [managed_node1] => {
"changed": false,
"rc": 0,
"results": []
}
MSG:
Nothing to do
TASK [fedora.linux_system_roles.ha_cluster : Check and prepare role variables] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:17
Monday 29 July 2024 08:18:01 -0400 (0:00:02.601) 0:00:07.317 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Discover cluster node names] ******
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:3
Monday 29 July 2024 08:18:01 -0400 (0:00:00.048) 0:00:07.365 ***********
ok: [managed_node1] => {
"ansible_facts": {
"__ha_cluster_node_name": "localhost"
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Collect cluster node names] *******
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:7
Monday 29 July 2024 08:18:01 -0400 (0:00:00.031) 0:00:07.396 ***********
ok: [managed_node1] => {
"ansible_facts": {
"__ha_cluster_all_node_names": [
"localhost"
]
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if ha_cluster_node_options contains unknown or duplicate nodes] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:16
Monday 29 July 2024 08:18:01 -0400 (0:00:00.034) 0:00:07.431 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Extract node options] *************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:30
Monday 29 July 2024 08:18:01 -0400 (0:00:00.029) 0:00:07.461 ***********
ok: [managed_node1] => {
"ansible_facts": {
"__ha_cluster_local_node": {}
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if passwords are not specified] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:43
Monday 29 July 2024 08:18:01 -0400 (0:00:00.033) 0:00:07.495 ***********
skipping: [managed_node1] => (item=ha_cluster_hacluster_password) => {
"ansible_loop_var": "item",
"changed": false,
"item": "ha_cluster_hacluster_password",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if nodes do not have the same number of SBD devices specified] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:53
Monday 29 July 2024 08:18:01 -0400 (0:00:00.053) 0:00:07.548 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if configuring qnetd on a cluster node] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:69
Monday 29 July 2024 08:18:01 -0400 (0:00:00.033) 0:00:07.582 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if no valid level is specified for a fencing level] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:79
Monday 29 July 2024 08:18:02 -0400 (0:00:00.029) 0:00:07.612 ***********
TASK [fedora.linux_system_roles.ha_cluster : Fail if no target is specified for a fencing level] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:87
Monday 29 July 2024 08:18:02 -0400 (0:00:00.024) 0:00:07.637 ***********
TASK [fedora.linux_system_roles.ha_cluster : Extract qdevice settings] *********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:101
Monday 29 July 2024 08:18:02 -0400 (0:00:00.025) 0:00:07.662 ***********
ok: [managed_node1] => {
"ansible_facts": {
"__ha_cluster_qdevice_host": "",
"__ha_cluster_qdevice_in_use": false,
"__ha_cluster_qdevice_model": "",
"__ha_cluster_qdevice_pcs_address": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Figure out if ATB needs to be enabled for SBD] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:110
Monday 29 July 2024 08:18:02 -0400 (0:00:00.036) 0:00:07.698 ***********
ok: [managed_node1] => {
"ansible_facts": {
"__ha_cluster_sbd_needs_atb": false
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if SBD needs ATB enabled and the user configured ATB to be disabled] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:120
Monday 29 July 2024 08:18:02 -0400 (0:00:00.032) 0:00:07.730 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if ha_cluster_pcsd_public_key_src and ha_cluster_pcsd_private_key_src are set along with ha_cluster_pcsd_certificates] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:127
Monday 29 July 2024 08:18:02 -0400 (0:00:00.027) 0:00:07.758 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch pcs capabilities] ***********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:141
Monday 29 July 2024 08:18:02 -0400 (0:00:00.027) 0:00:07.785 ***********
ok: [managed_node1] => {
"changed": false,
"cmd": [
"pcs",
"--version",
"--full"
],
"delta": "0:00:00.502590",
"end": "2024-07-29 08:18:02.987564",
"rc": 0,
"start": "2024-07-29 08:18:02.484974"
}
STDOUT:
0.10.18
booth booth.enable-authfile.set booth.enable-authfile.unset cluster.config.backup-local cluster.config.restore-cluster cluster.config.restore-local cluster.config.uuid cluster.create cluster.create.enable cluster.create.local cluster.create.no-keys-sync cluster.create.separated-name-and-address cluster.create.start cluster.create.start.wait cluster.create.transport.knet cluster.create.transport.udp-udpu cluster.create.transport.udp-udpu.no-rrp cluster.destroy cluster.destroy.all cluster.report cluster.verify corosync.authkey.update corosync.config.get corosync.config.get.struct corosync.config.reload corosync.config.sync-to-local-cluster corosync.config.update corosync.link.add corosync.link.remove corosync.link.remove.list corosync.link.update corosync.qdevice corosync.qdevice.model.net corosync.quorum corosync.quorum.device corosync.quorum.device.heuristics corosync.quorum.device.model.net corosync.quorum.device.model.net.options_tls_and_kaptb corosync.quorum.set-expected-votes-runtime corosync.quorum.status corosync.quorum.unblock corosync.totem.block_unlisted_ips corosync.uidgid node.add node.add.enable node.add.separated-name-and-address node.add.start node.add.start.wait node.attributes node.attributes.set-list-for-node node.confirm-off node.fence node.guest node.kill node.maintenance node.maintenance.all node.maintenance.list node.maintenance.wait node.remote node.remote.onfail-demote node.remove node.remove-from-caches node.remove.list node.standby node.standby.all node.standby.list node.standby.wait node.start-stop-enable-disable node.start-stop-enable-disable.all node.start-stop-enable-disable.list node.start-stop-enable-disable.start-wait node.utilization node.utilization.set-list-for-node pcmk.acl.enable-disable pcmk.acl.group pcmk.acl.role pcmk.acl.role.create-with-permissions pcmk.acl.role.delete-with-users-groups pcmk.acl.user pcmk.alert pcmk.cib.checkpoints pcmk.cib.checkpoints.diff pcmk.cib.edit pcmk.cib.get pcmk.cib.get.scope pcmk.cib.roles.promoted-unpromoted pcmk.cib.set pcmk.constraint.colocation.set pcmk.constraint.colocation.set.options pcmk.constraint.colocation.simple pcmk.constraint.colocation.simple.options pcmk.constraint.hide-expired pcmk.constraint.location.simple pcmk.constraint.location.simple.options pcmk.constraint.location.simple.resource-regexp pcmk.constraint.location.simple.rule pcmk.constraint.location.simple.rule.node-attr-type-number pcmk.constraint.location.simple.rule.options pcmk.constraint.location.simple.rule.rule-add-remove pcmk.constraint.no-autocorrect pcmk.constraint.order.set pcmk.constraint.order.set.options pcmk.constraint.order.simple pcmk.constraint.order.simple.options pcmk.constraint.ticket.set pcmk.constraint.ticket.set.options pcmk.constraint.ticket.simple pcmk.constraint.ticket.simple.constraint-id pcmk.properties.cluster pcmk.properties.cluster.config.output-formats pcmk.properties.cluster.defaults pcmk.properties.cluster.describe pcmk.properties.cluster.describe.output-formats pcmk.properties.operation-defaults pcmk.properties.operation-defaults.multiple pcmk.properties.operation-defaults.rule pcmk.properties.operation-defaults.rule-rsc-op pcmk.properties.operation-defaults.rule.hide-expired pcmk.properties.operation-defaults.rule.node-attr-type-number pcmk.properties.resource-defaults pcmk.properties.resource-defaults.multiple pcmk.properties.resource-defaults.rule pcmk.properties.resource-defaults.rule-rsc-op pcmk.properties.resource-defaults.rule.hide-expired pcmk.properties.resource-defaults.rule.node-attr-type-number pcmk.resource.ban-move-clear pcmk.resource.ban-move-clear.bundles pcmk.resource.ban-move-clear.clear-expired pcmk.resource.ban-move-clear.clone pcmk.resource.bundle pcmk.resource.bundle.container-docker pcmk.resource.bundle.container-docker.promoted-max pcmk.resource.bundle.container-podman pcmk.resource.bundle.container-podman.promoted-max pcmk.resource.bundle.container-rkt pcmk.resource.bundle.container-rkt.promoted-max pcmk.resource.bundle.reset pcmk.resource.bundle.wait pcmk.resource.cleanup pcmk.resource.cleanup.one-resource pcmk.resource.cleanup.strict pcmk.resource.clone pcmk.resource.clone.custom-id pcmk.resource.clone.meta-in-create pcmk.resource.clone.wait pcmk.resource.config.output-formats pcmk.resource.create pcmk.resource.create.clone.custom-id pcmk.resource.create.in-existing-bundle pcmk.resource.create.meta pcmk.resource.create.no-master pcmk.resource.create.operations pcmk.resource.create.operations.onfail-demote pcmk.resource.create.promotable pcmk.resource.create.promotable.custom-id pcmk.resource.create.wait pcmk.resource.debug pcmk.resource.delete pcmk.resource.disable.safe pcmk.resource.disable.safe.brief pcmk.resource.disable.safe.tag pcmk.resource.disable.simulate pcmk.resource.disable.simulate.brief pcmk.resource.disable.simulate.tag pcmk.resource.enable-disable pcmk.resource.enable-disable.list pcmk.resource.enable-disable.tag pcmk.resource.enable-disable.wait pcmk.resource.failcount pcmk.resource.group pcmk.resource.group.add-remove-list pcmk.resource.group.wait pcmk.resource.manage-unmanage pcmk.resource.manage-unmanage.list pcmk.resource.manage-unmanage.tag pcmk.resource.manage-unmanage.with-monitor pcmk.resource.move.autoclean pcmk.resource.promotable pcmk.resource.promotable.custom-id pcmk.resource.promotable.meta-in-create pcmk.resource.promotable.wait pcmk.resource.refresh pcmk.resource.refresh.one-resource pcmk.resource.refresh.strict pcmk.resource.relations pcmk.resource.relocate pcmk.resource.restart pcmk.resource.update pcmk.resource.update-meta pcmk.resource.update-meta.list pcmk.resource.update-meta.wait pcmk.resource.update-operations pcmk.resource.update-operations.onfail-demote pcmk.resource.update.meta pcmk.resource.update.operations pcmk.resource.update.operations.onfail-demote pcmk.resource.update.wait pcmk.resource.utilization pcmk.resource.utilization-set-list-for-resource pcmk.stonith.cleanup pcmk.stonith.cleanup.one-resource pcmk.stonith.cleanup.strict pcmk.stonith.create pcmk.stonith.create.in-group pcmk.stonith.create.meta pcmk.stonith.create.operations pcmk.stonith.create.operations.onfail-demote pcmk.stonith.create.wait pcmk.stonith.delete pcmk.stonith.enable-disable pcmk.stonith.enable-disable.list pcmk.stonith.enable-disable.wait pcmk.stonith.history.cleanup pcmk.stonith.history.show pcmk.stonith.history.update pcmk.stonith.levels pcmk.stonith.levels.add-remove-devices-list pcmk.stonith.levels.clear pcmk.stonith.levels.node-attr pcmk.stonith.levels.node-regexp pcmk.stonith.levels.verify pcmk.stonith.refresh pcmk.stonith.refresh.one-resource pcmk.stonith.refresh.strict pcmk.stonith.update pcmk.stonith.update.scsi-devices pcmk.stonith.update.scsi-devices.add-remove pcmk.stonith.update.scsi-devices.mpath pcmk.tag pcmk.tag.resources pcs.auth.client pcs.auth.client.cluster pcs.auth.client.token pcs.auth.deauth-client pcs.auth.deauth-server pcs.auth.no-bidirectional pcs.auth.separated-name-and-address pcs.auth.server.token pcs.cfg-in-file.cib pcs.daemon-ssl-cert.set pcs.daemon-ssl-cert.sync-to-local-cluster pcs.disaster-recovery.essentials pcs.request-timeout resource-agents.describe resource-agents.list resource-agents.list.detailed resource-agents.ocf.version-1-0 resource-agents.ocf.version-1-1 resource-agents.self-validation sbd sbd.option-timeout-action sbd.shared-block-device status.corosync.membership status.pcmk.resources.hide-inactive status.pcmk.resources.id status.pcmk.resources.node status.pcmk.resources.orphaned status.pcmk.xml stonith-agents.describe stonith-agents.list stonith-agents.list.detailed stonith-agents.ocf.version-1-0 stonith-agents.ocf.version-1-1 stonith-agents.self-validation
TASK [fedora.linux_system_roles.ha_cluster : Parse pcs capabilities] ***********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:148
Monday 29 July 2024 08:18:03 -0400 (0:00:00.840) 0:00:08.626 ***********
ok: [managed_node1] => {
"ansible_facts": {
"__ha_cluster_pcs_capabilities": [
"booth",
"booth.enable-authfile.set",
"booth.enable-authfile.unset",
"cluster.config.backup-local",
"cluster.config.restore-cluster",
"cluster.config.restore-local",
"cluster.config.uuid",
"cluster.create",
"cluster.create.enable",
"cluster.create.local",
"cluster.create.no-keys-sync",
"cluster.create.separated-name-and-address",
"cluster.create.start",
"cluster.create.start.wait",
"cluster.create.transport.knet",
"cluster.create.transport.udp-udpu",
"cluster.create.transport.udp-udpu.no-rrp",
"cluster.destroy",
"cluster.destroy.all",
"cluster.report",
"cluster.verify",
"corosync.authkey.update",
"corosync.config.get",
"corosync.config.get.struct",
"corosync.config.reload",
"corosync.config.sync-to-local-cluster",
"corosync.config.update",
"corosync.link.add",
"corosync.link.remove",
"corosync.link.remove.list",
"corosync.link.update",
"corosync.qdevice",
"corosync.qdevice.model.net",
"corosync.quorum",
"corosync.quorum.device",
"corosync.quorum.device.heuristics",
"corosync.quorum.device.model.net",
"corosync.quorum.device.model.net.options_tls_and_kaptb",
"corosync.quorum.set-expected-votes-runtime",
"corosync.quorum.status",
"corosync.quorum.unblock",
"corosync.totem.block_unlisted_ips",
"corosync.uidgid",
"node.add",
"node.add.enable",
"node.add.separated-name-and-address",
"node.add.start",
"node.add.start.wait",
"node.attributes",
"node.attributes.set-list-for-node",
"node.confirm-off",
"node.fence",
"node.guest",
"node.kill",
"node.maintenance",
"node.maintenance.all",
"node.maintenance.list",
"node.maintenance.wait",
"node.remote",
"node.remote.onfail-demote",
"node.remove",
"node.remove-from-caches",
"node.remove.list",
"node.standby",
"node.standby.all",
"node.standby.list",
"node.standby.wait",
"node.start-stop-enable-disable",
"node.start-stop-enable-disable.all",
"node.start-stop-enable-disable.list",
"node.start-stop-enable-disable.start-wait",
"node.utilization",
"node.utilization.set-list-for-node",
"pcmk.acl.enable-disable",
"pcmk.acl.group",
"pcmk.acl.role",
"pcmk.acl.role.create-with-permissions",
"pcmk.acl.role.delete-with-users-groups",
"pcmk.acl.user",
"pcmk.alert",
"pcmk.cib.checkpoints",
"pcmk.cib.checkpoints.diff",
"pcmk.cib.edit",
"pcmk.cib.get",
"pcmk.cib.get.scope",
"pcmk.cib.roles.promoted-unpromoted",
"pcmk.cib.set",
"pcmk.constraint.colocation.set",
"pcmk.constraint.colocation.set.options",
"pcmk.constraint.colocation.simple",
"pcmk.constraint.colocation.simple.options",
"pcmk.constraint.hide-expired",
"pcmk.constraint.location.simple",
"pcmk.constraint.location.simple.options",
"pcmk.constraint.location.simple.resource-regexp",
"pcmk.constraint.location.simple.rule",
"pcmk.constraint.location.simple.rule.node-attr-type-number",
"pcmk.constraint.location.simple.rule.options",
"pcmk.constraint.location.simple.rule.rule-add-remove",
"pcmk.constraint.no-autocorrect",
"pcmk.constraint.order.set",
"pcmk.constraint.order.set.options",
"pcmk.constraint.order.simple",
"pcmk.constraint.order.simple.options",
"pcmk.constraint.ticket.set",
"pcmk.constraint.ticket.set.options",
"pcmk.constraint.ticket.simple",
"pcmk.constraint.ticket.simple.constraint-id",
"pcmk.properties.cluster",
"pcmk.properties.cluster.config.output-formats",
"pcmk.properties.cluster.defaults",
"pcmk.properties.cluster.describe",
"pcmk.properties.cluster.describe.output-formats",
"pcmk.properties.operation-defaults",
"pcmk.properties.operation-defaults.multiple",
"pcmk.properties.operation-defaults.rule",
"pcmk.properties.operation-defaults.rule-rsc-op",
"pcmk.properties.operation-defaults.rule.hide-expired",
"pcmk.properties.operation-defaults.rule.node-attr-type-number",
"pcmk.properties.resource-defaults",
"pcmk.properties.resource-defaults.multiple",
"pcmk.properties.resource-defaults.rule",
"pcmk.properties.resource-defaults.rule-rsc-op",
"pcmk.properties.resource-defaults.rule.hide-expired",
"pcmk.properties.resource-defaults.rule.node-attr-type-number",
"pcmk.resource.ban-move-clear",
"pcmk.resource.ban-move-clear.bundles",
"pcmk.resource.ban-move-clear.clear-expired",
"pcmk.resource.ban-move-clear.clone",
"pcmk.resource.bundle",
"pcmk.resource.bundle.container-docker",
"pcmk.resource.bundle.container-docker.promoted-max",
"pcmk.resource.bundle.container-podman",
"pcmk.resource.bundle.container-podman.promoted-max",
"pcmk.resource.bundle.container-rkt",
"pcmk.resource.bundle.container-rkt.promoted-max",
"pcmk.resource.bundle.reset",
"pcmk.resource.bundle.wait",
"pcmk.resource.cleanup",
"pcmk.resource.cleanup.one-resource",
"pcmk.resource.cleanup.strict",
"pcmk.resource.clone",
"pcmk.resource.clone.custom-id",
"pcmk.resource.clone.meta-in-create",
"pcmk.resource.clone.wait",
"pcmk.resource.config.output-formats",
"pcmk.resource.create",
"pcmk.resource.create.clone.custom-id",
"pcmk.resource.create.in-existing-bundle",
"pcmk.resource.create.meta",
"pcmk.resource.create.no-master",
"pcmk.resource.create.operations",
"pcmk.resource.create.operations.onfail-demote",
"pcmk.resource.create.promotable",
"pcmk.resource.create.promotable.custom-id",
"pcmk.resource.create.wait",
"pcmk.resource.debug",
"pcmk.resource.delete",
"pcmk.resource.disable.safe",
"pcmk.resource.disable.safe.brief",
"pcmk.resource.disable.safe.tag",
"pcmk.resource.disable.simulate",
"pcmk.resource.disable.simulate.brief",
"pcmk.resource.disable.simulate.tag",
"pcmk.resource.enable-disable",
"pcmk.resource.enable-disable.list",
"pcmk.resource.enable-disable.tag",
"pcmk.resource.enable-disable.wait",
"pcmk.resource.failcount",
"pcmk.resource.group",
"pcmk.resource.group.add-remove-list",
"pcmk.resource.group.wait",
"pcmk.resource.manage-unmanage",
"pcmk.resource.manage-unmanage.list",
"pcmk.resource.manage-unmanage.tag",
"pcmk.resource.manage-unmanage.with-monitor",
"pcmk.resource.move.autoclean",
"pcmk.resource.promotable",
"pcmk.resource.promotable.custom-id",
"pcmk.resource.promotable.meta-in-create",
"pcmk.resource.promotable.wait",
"pcmk.resource.refresh",
"pcmk.resource.refresh.one-resource",
"pcmk.resource.refresh.strict",
"pcmk.resource.relations",
"pcmk.resource.relocate",
"pcmk.resource.restart",
"pcmk.resource.update",
"pcmk.resource.update-meta",
"pcmk.resource.update-meta.list",
"pcmk.resource.update-meta.wait",
"pcmk.resource.update-operations",
"pcmk.resource.update-operations.onfail-demote",
"pcmk.resource.update.meta",
"pcmk.resource.update.operations",
"pcmk.resource.update.operations.onfail-demote",
"pcmk.resource.update.wait",
"pcmk.resource.utilization",
"pcmk.resource.utilization-set-list-for-resource",
"pcmk.stonith.cleanup",
"pcmk.stonith.cleanup.one-resource",
"pcmk.stonith.cleanup.strict",
"pcmk.stonith.create",
"pcmk.stonith.create.in-group",
"pcmk.stonith.create.meta",
"pcmk.stonith.create.operations",
"pcmk.stonith.create.operations.onfail-demote",
"pcmk.stonith.create.wait",
"pcmk.stonith.delete",
"pcmk.stonith.enable-disable",
"pcmk.stonith.enable-disable.list",
"pcmk.stonith.enable-disable.wait",
"pcmk.stonith.history.cleanup",
"pcmk.stonith.history.show",
"pcmk.stonith.history.update",
"pcmk.stonith.levels",
"pcmk.stonith.levels.add-remove-devices-list",
"pcmk.stonith.levels.clear",
"pcmk.stonith.levels.node-attr",
"pcmk.stonith.levels.node-regexp",
"pcmk.stonith.levels.verify",
"pcmk.stonith.refresh",
"pcmk.stonith.refresh.one-resource",
"pcmk.stonith.refresh.strict",
"pcmk.stonith.update",
"pcmk.stonith.update.scsi-devices",
"pcmk.stonith.update.scsi-devices.add-remove",
"pcmk.stonith.update.scsi-devices.mpath",
"pcmk.tag",
"pcmk.tag.resources",
"pcs.auth.client",
"pcs.auth.client.cluster",
"pcs.auth.client.token",
"pcs.auth.deauth-client",
"pcs.auth.deauth-server",
"pcs.auth.no-bidirectional",
"pcs.auth.separated-name-and-address",
"pcs.auth.server.token",
"pcs.cfg-in-file.cib",
"pcs.daemon-ssl-cert.set",
"pcs.daemon-ssl-cert.sync-to-local-cluster",
"pcs.disaster-recovery.essentials",
"pcs.request-timeout",
"resource-agents.describe",
"resource-agents.list",
"resource-agents.list.detailed",
"resource-agents.ocf.version-1-0",
"resource-agents.ocf.version-1-1",
"resource-agents.self-validation",
"sbd",
"sbd.option-timeout-action",
"sbd.shared-block-device",
"status.corosync.membership",
"status.pcmk.resources.hide-inactive",
"status.pcmk.resources.id",
"status.pcmk.resources.node",
"status.pcmk.resources.orphaned",
"status.pcmk.xml",
"stonith-agents.describe",
"stonith-agents.list",
"stonith-agents.list.detailed",
"stonith-agents.ocf.version-1-0",
"stonith-agents.ocf.version-1-1",
"stonith-agents.self-validation"
],
"__ha_cluster_pcsd_capabilities_available": false
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch pcsd capabilities] **********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:155
Monday 29 July 2024 08:18:03 -0400 (0:00:00.035) 0:00:08.662 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Parse pcsd capabilities] **********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:163
Monday 29 July 2024 08:18:03 -0400 (0:00:00.010) 0:00:08.672 ***********
ok: [managed_node1] => {
"ansible_facts": {
"__ha_cluster_pcsd_capabilities": []
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if pcs is to old to configure resources and operations defaults] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:172
Monday 29 July 2024 08:18:03 -0400 (0:00:00.030) 0:00:08.703 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Set hacluster password] ***********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:22
Monday 29 July 2024 08:18:03 -0400 (0:00:00.027) 0:00:08.731 ***********
ok: [managed_node1] => {
"append": false,
"changed": false,
"comment": "cluster user",
"group": 189,
"home": "/home/hacluster",
"move_home": false,
"name": "hacluster",
"password": "NOT_LOGGING_PASSWORD",
"shell": "/sbin/nologin",
"state": "present",
"uid": 189
}
TASK [fedora.linux_system_roles.ha_cluster : Configure shell] ******************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:29
Monday 29 July 2024 08:18:03 -0400 (0:00:00.556) 0:00:09.287 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Stop pcsd] ************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:6
Monday 29 July 2024 08:18:03 -0400 (0:00:00.070) 0:00:09.358 ***********
changed: [managed_node1] => {
"changed": true,
"name": "pcsd",
"state": "stopped",
"status": {
"ActiveEnterTimestamp": "Mon 2024-07-29 08:15:39 EDT",
"ActiveEnterTimestampMonotonic": "372269177",
"ActiveExitTimestamp": "Mon 2024-07-29 08:15:37 EDT",
"ActiveExitTimestampMonotonic": "370105608",
"ActiveState": "active",
"After": "pcsd-ruby.service network-online.target basic.target sysinit.target system.slice systemd-journald.socket",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Mon 2024-07-29 08:15:39 EDT",
"AssertTimestampMonotonic": "372029023",
"Before": "shutdown.target multi-user.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Mon 2024-07-29 08:15:39 EDT",
"ConditionTimestampMonotonic": "372028999",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ConsistsOf": "pcsd-ruby.service",
"ControlGroup": "/system.slice/pcsd.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "PCS GUI and remote configuration interface",
"DevicePolicy": "auto",
"Documentation": "man:pcsd(8) man:pcs(8)",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/pcsd (ignore_errors=no)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "26433",
"ExecMainStartTimestamp": "Mon 2024-07-29 08:15:39 EDT",
"ExecMainStartTimestampMonotonic": "372030271",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/pcsd ; argv[]=/usr/sbin/pcsd ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/pcsd.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "pcsd.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestamp": "Mon 2024-07-29 08:15:37 EDT",
"InactiveEnterTimestampMonotonic": "370143133",
"InactiveExitTimestamp": "Mon 2024-07-29 08:15:39 EDT",
"InactiveExitTimestampMonotonic": "372030334",
"InvocationID": "b49c46f5442a4b7d90474730797d1db2",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14004",
"LimitNPROCSoft": "14004",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14004",
"LimitSIGPENDINGSoft": "14004",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "26433",
"MemoryAccounting": "yes",
"MemoryCurrent": "23658496",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "pcsd.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "network-online.target sysinit.target system.slice pcsd-ruby.service",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Mon 2024-07-29 08:15:39 EDT",
"StateChangeTimestampMonotonic": "372269177",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "1",
"TasksMax": "22406",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogTimestamp": "Mon 2024-07-29 08:15:39 EDT",
"WatchdogTimestampMonotonic": "372269173",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Regenerate pcsd TLS certificate and key] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:11
Monday 29 July 2024 08:18:04 -0400 (0:00:00.708) 0:00:10.067 ***********
skipping: [managed_node1] => (item=/var/lib/pcsd/pcsd.key) => {
"ansible_loop_var": "item",
"changed": false,
"item": "/var/lib/pcsd/pcsd.key",
"skip_reason": "Conditional result was False"
}
skipping: [managed_node1] => (item=/var/lib/pcsd/pcsd.crt) => {
"ansible_loop_var": "item",
"changed": false,
"item": "/var/lib/pcsd/pcsd.crt",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Get the stat of /var/lib/pcsd] ****
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:25
Monday 29 July 2024 08:18:04 -0400 (0:00:00.028) 0:00:10.095 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Allow certmonger to write into pcsd's certificate directory] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:30
Monday 29 July 2024 08:18:04 -0400 (0:00:00.030) 0:00:10.125 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Ensure the name of ha_cluster_pcsd_certificates is /var/lib/pcsd/pcsd; Create certificates using the certificate role] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:37
Monday 29 July 2024 08:18:04 -0400 (0:00:00.027) 0:00:10.153 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Set pcsd's certificate directory back to cluster_var_lib_t] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:49
Monday 29 July 2024 08:18:04 -0400 (0:00:00.027) 0:00:10.181 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute pcsd TLS private key] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:64
Monday 29 July 2024 08:18:04 -0400 (0:00:00.029) 0:00:10.210 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute pcsd TLS certificate] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:71
Monday 29 July 2024 08:18:04 -0400 (0:00:00.028) 0:00:10.239 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute pcs_settings.conf] *****
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:79
Monday 29 July 2024 08:18:04 -0400 (0:00:00.033) 0:00:10.273 ***********
ok: [managed_node1] => {
"changed": false,
"checksum": "9bbea634a798cf0976b80cd3c4e34aca6a6a7d44",
"dest": "/var/lib/pcsd/pcs_settings.conf",
"gid": 0,
"group": "root",
"mode": "0644",
"owner": "root",
"path": "/var/lib/pcsd/pcs_settings.conf",
"secontext": "system_u:object_r:cluster_var_lib_t:s0",
"size": 361,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Start pcsd with updated config files and configure it to start on boot] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:88
Monday 29 July 2024 08:18:05 -0400 (0:00:00.613) 0:00:10.886 ***********
changed: [managed_node1] => {
"changed": true,
"enabled": true,
"name": "pcsd",
"state": "started",
"status": {
"ActiveEnterTimestamp": "Mon 2024-07-29 08:15:39 EDT",
"ActiveEnterTimestampMonotonic": "372269177",
"ActiveExitTimestamp": "Mon 2024-07-29 08:18:04 EDT",
"ActiveExitTimestampMonotonic": "517390587",
"ActiveState": "inactive",
"After": "pcsd-ruby.service network-online.target basic.target sysinit.target system.slice systemd-journald.socket",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Mon 2024-07-29 08:15:39 EDT",
"AssertTimestampMonotonic": "372029023",
"Before": "shutdown.target multi-user.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Mon 2024-07-29 08:15:39 EDT",
"ConditionTimestampMonotonic": "372028999",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ConsistsOf": "pcsd-ruby.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "PCS GUI and remote configuration interface",
"DevicePolicy": "auto",
"Documentation": "man:pcsd(8) man:pcs(8)",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/pcsd (ignore_errors=no)",
"ExecMainCode": "1",
"ExecMainExitTimestamp": "Mon 2024-07-29 08:18:04 EDT",
"ExecMainExitTimestampMonotonic": "517420934",
"ExecMainPID": "26433",
"ExecMainStartTimestamp": "Mon 2024-07-29 08:15:39 EDT",
"ExecMainStartTimestampMonotonic": "372030271",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/pcsd ; argv[]=/usr/sbin/pcsd ; ignore_errors=no ; start_time=[Mon 2024-07-29 08:15:39 EDT] ; stop_time=[Mon 2024-07-29 08:18:04 EDT] ; pid=26433 ; code=exited ; status=0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/pcsd.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "pcsd.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestamp": "Mon 2024-07-29 08:18:04 EDT",
"InactiveEnterTimestampMonotonic": "517421065",
"InactiveExitTimestamp": "Mon 2024-07-29 08:15:39 EDT",
"InactiveExitTimestampMonotonic": "372030334",
"InvocationID": "b49c46f5442a4b7d90474730797d1db2",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14004",
"LimitNPROCSoft": "14004",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14004",
"LimitSIGPENDINGSoft": "14004",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "pcsd.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "network-online.target sysinit.target system.slice pcsd-ruby.service",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Mon 2024-07-29 08:18:04 EDT",
"StateChangeTimestampMonotonic": "517421065",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "22406",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Configure firewall] ***************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:35
Monday 29 July 2024 08:18:06 -0400 (0:00:01.152) 0:00:12.038 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/firewall.yml for managed_node1
TASK [Ensure the service and the ports status with the firewall role] **********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/firewall.yml:3
Monday 29 July 2024 08:18:06 -0400 (0:00:00.036) 0:00:12.074 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure selinux] ****************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:38
Monday 29 July 2024 08:18:06 -0400 (0:00:00.030) 0:00:12.104 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Populate service facts] ***********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml:3
Monday 29 July 2024 08:18:06 -0400 (0:00:00.038) 0:00:12.143 ***********
ok: [managed_node1] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"corosync-notifyd.service": {
"name": "corosync-notifyd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qnetd.service": {
"name": "corosync-qnetd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync.service": {
"name": "corosync.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crm_mon.service": {
"name": "crm_mon.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "unknown"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-pvscan@.service": {
"name": "lvm2-pvscan@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"multipathd.service": {
"name": "multipathd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"pacemaker.service": {
"name": "pacemaker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"pcsd-ruby.service": {
"name": "pcsd-ruby.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"pcsd.service": {
"name": "pcsd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sbd.service": {
"name": "sbd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"sbd_remote.service": {
"name": "sbd_remote.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "unknown"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Set the fence-virt/fence-agents port to _ha_cluster_selinux] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml:12
Monday 29 July 2024 08:18:08 -0400 (0:00:01.685) 0:00:13.828 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Get associated selinux ports] *****
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml:20
Monday 29 July 2024 08:18:08 -0400 (0:00:00.038) 0:00:13.867 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Add the high-availability service ports to _ha_cluster_selinux] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml:28
Monday 29 July 2024 08:18:08 -0400 (0:00:00.036) 0:00:13.904 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Ensure the service and the ports status with the selinux role] ***********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml:35
Monday 29 July 2024 08:18:08 -0400 (0:00:00.040) 0:00:13.944 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Install cluster packages] *********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:44
Monday 29 July 2024 08:18:08 -0400 (0:00:00.039) 0:00:13.983 ***********
ok: [managed_node1] => {
"changed": false,
"rc": 0,
"results": []
}
MSG:
Nothing to do
TASK [fedora.linux_system_roles.ha_cluster : Distribute fence-virt authkey] ****
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:50
Monday 29 July 2024 08:18:10 -0400 (0:00:02.486) 0:00:16.469 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/distribute-fence-virt-key.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Create /etc/cluster directory] ****
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/distribute-fence-virt-key.yml:3
Monday 29 July 2024 08:18:10 -0400 (0:00:00.043) 0:00:16.512 ***********
ok: [managed_node1] => {
"changed": false,
"gid": 0,
"group": "root",
"mode": "0755",
"owner": "root",
"path": "/etc/cluster",
"secontext": "unconfined_u:object_r:etc_t:s0",
"size": 27,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Get fence_xvm.key] ****************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/distribute-fence-virt-key.yml:9
Monday 29 July 2024 08:18:11 -0400 (0:00:00.371) 0:00:16.884 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Check if key exists on the controller fence_xvm.key] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:16
Monday 29 July 2024 08:18:11 -0400 (0:00:00.060) 0:00:16.944 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Slurp existing key from the controller fence_xvm.key] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:21
Monday 29 July 2024 08:18:11 -0400 (0:00:00.022) 0:00:16.967 ***********
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Use the slurped key from the controller fence_xvm.key] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:29
Monday 29 July 2024 08:18:11 -0400 (0:00:00.021) 0:00:16.989 ***********
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Generate key using OpenSSL fence_xvm.key] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:44
Monday 29 July 2024 08:18:11 -0400 (0:00:00.020) 0:00:17.010 ***********
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch generated fence_xvm.key] ****
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:50
Monday 29 July 2024 08:18:11 -0400 (0:00:00.357) 0:00:17.368 ***********
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Check if key exists on cluster nodes fence_xvm.key] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:60
Monday 29 July 2024 08:18:11 -0400 (0:00:00.026) 0:00:17.395 ***********
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1722255364.347214,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 8,
"charset": "binary",
"checksum": "d9d3a00c278df4c528d31bd06966b3c579f276ee",
"ctime": 1722255258.5853577,
"dev": 51713,
"device_type": 0,
"executable": false,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 50331866,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/octet-stream",
"mode": "0600",
"mtime": 1722255258.312358,
"nlink": 1,
"path": "/etc/cluster/fence_xvm.key",
"pw_name": "root",
"readable": true,
"rgrp": false,
"roth": false,
"rusr": true,
"size": 512,
"uid": 0,
"version": "1268902098",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": false,
"xoth": false,
"xusr": false
}
}
TASK [fedora.linux_system_roles.ha_cluster : Slurp existing key from cluster nodes fence_xvm.key] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:65
Monday 29 July 2024 08:18:12 -0400 (0:00:00.356) 0:00:17.751 ***********
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Use the slurped key from cluster nodes fence_xvm.key] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:73
Monday 29 July 2024 08:18:12 -0400 (0:00:00.451) 0:00:18.202 ***********
ok: [managed_node1 -> managed_node1] => (item=None) => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
ok: [managed_node1 -> {{ item }}] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Send fence_xvm.key to nodes] ******
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/distribute-fence-virt-key.yml:17
Monday 29 July 2024 08:18:12 -0400 (0:00:00.035) 0:00:18.238 ***********
ok: [managed_node1] => {
"changed": false,
"checksum": "d9d3a00c278df4c528d31bd06966b3c579f276ee",
"dest": "/etc/cluster/fence_xvm.key",
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/etc/cluster/fence_xvm.key",
"secontext": "system_u:object_r:cluster_conf_t:s0",
"size": 512,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Configure SBD] ********************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:55
Monday 29 July 2024 08:18:13 -0400 (0:00:00.602) 0:00:18.840 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Configure watchdog kernel module blocklist] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:10
Monday 29 July 2024 08:18:13 -0400 (0:00:00.045) 0:00:18.885 ***********
TASK [fedora.linux_system_roles.ha_cluster : Unload watchdog kernel modules from blocklist] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:22
Monday 29 July 2024 08:18:13 -0400 (0:00:00.018) 0:00:18.904 ***********
TASK [fedora.linux_system_roles.ha_cluster : Configure watchdog kernel modules] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:30
Monday 29 July 2024 08:18:13 -0400 (0:00:00.016) 0:00:18.921 ***********
TASK [fedora.linux_system_roles.ha_cluster : Load watchdog kernel modules] *****
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:40
Monday 29 July 2024 08:18:13 -0400 (0:00:00.016) 0:00:18.937 ***********
TASK [fedora.linux_system_roles.ha_cluster : Probe SBD devices] ****************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:57
Monday 29 July 2024 08:18:13 -0400 (0:00:00.018) 0:00:18.956 ***********
TASK [fedora.linux_system_roles.ha_cluster : Initialize SBD devices] ***********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:70
Monday 29 July 2024 08:18:13 -0400 (0:00:00.016) 0:00:18.973 ***********
TASK [fedora.linux_system_roles.ha_cluster : Distribute SBD config] ************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:88
Monday 29 July 2024 08:18:13 -0400 (0:00:00.015) 0:00:18.988 ***********
changed: [managed_node1] => {
"changed": true,
"checksum": "d5aa0d6d6a774ebaa2eb634f5ef174bb0688c5f8",
"dest": "/etc/sysconfig/sbd",
"gid": 0,
"group": "root",
"md5sum": "6136a912bc0ec0daff3f97cd6fbd3985",
"mode": "0644",
"owner": "root",
"secontext": "system_u:object_r:etc_t:s0",
"size": 4271,
"src": "/root/.ansible/tmp/ansible-tmp-1722255493.4463773-11809-153806655576635/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Ensure /etc/systemd/system/sbd.service.d directory exists] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:109
Monday 29 July 2024 08:18:14 -0400 (0:00:00.766) 0:00:19.754 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Override start timeout for SBD] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:117
Monday 29 July 2024 08:18:14 -0400 (0:00:00.021) 0:00:19.776 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Reload systemd service files] *****
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:131
Monday 29 July 2024 08:18:14 -0400 (0:00:00.024) 0:00:19.801 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Get services status - detect pacemaker] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:135
Monday 29 July 2024 08:18:14 -0400 (0:00:00.021) 0:00:19.823 ***********
ok: [managed_node1] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"corosync-notifyd.service": {
"name": "corosync-notifyd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qnetd.service": {
"name": "corosync-qnetd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync.service": {
"name": "corosync.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crm_mon.service": {
"name": "crm_mon.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "unknown"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-pvscan@.service": {
"name": "lvm2-pvscan@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"multipathd.service": {
"name": "multipathd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"pacemaker.service": {
"name": "pacemaker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"pcsd-ruby.service": {
"name": "pcsd-ruby.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"pcsd.service": {
"name": "pcsd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sbd.service": {
"name": "sbd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"sbd_remote.service": {
"name": "sbd_remote.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "unknown"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Set stonith-watchdog-timeout cluster property in CIB] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:142
Monday 29 July 2024 08:18:15 -0400 (0:00:01.554) 0:00:21.378 ***********
changed: [managed_node1] => {
"changed": true,
"cmd": [
"pcs",
"--force",
"--",
"property",
"set",
"stonith-watchdog-timeout="
],
"delta": "0:00:00.584504",
"end": "2024-07-29 08:18:16.672360",
"rc": 0,
"start": "2024-07-29 08:18:16.087856"
}
TASK [fedora.linux_system_roles.ha_cluster : Correct cib.xml ownership] ********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:155
Monday 29 July 2024 08:18:16 -0400 (0:00:00.934) 0:00:22.313 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Clean cib.xml.sig] ****************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:167
Monday 29 July 2024 08:18:16 -0400 (0:00:00.022) 0:00:22.335 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure corosync] ***************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:58
Monday 29 July 2024 08:18:16 -0400 (0:00:00.021) 0:00:22.357 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-corosync.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Create a corosync.conf tempfile] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-corosync.yml:3
Monday 29 July 2024 08:18:16 -0400 (0:00:00.039) 0:00:22.396 ***********
changed: [managed_node1] => {
"changed": true,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.eyt1mp6k_ha_cluster_corosync_conf",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Create a corosync.conf file content] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-corosync.yml:16
Monday 29 July 2024 08:18:17 -0400 (0:00:00.340) 0:00:22.736 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cluster-setup-pcs-0.10.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Create a corosync.conf file content using pcs-0.10] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cluster-setup-pcs-0.10.yml:3
Monday 29 July 2024 08:18:17 -0400 (0:00:00.029) 0:00:22.766 ***********
changed: [managed_node1] => {
"changed": true,
"cmd": [
"pcs",
"cluster",
"setup",
"--corosync_conf",
"/tmp/ansible.eyt1mp6k_ha_cluster_corosync_conf",
"--overwrite",
"--no-cluster-uuid",
"--",
"test-cluster",
"localhost"
],
"delta": "0:00:00.493320",
"end": "2024-07-29 08:18:17.984652",
"rc": 0,
"start": "2024-07-29 08:18:17.491332"
}
STDOUT:
No addresses specified for host 'localhost', using 'localhost'
TASK [fedora.linux_system_roles.ha_cluster : Add qdevice configuration to corosync.conf] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cluster-setup-pcs-0.10.yml:81
Monday 29 July 2024 08:18:18 -0400 (0:00:00.862) 0:00:23.628 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch created corosync.conf file] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-corosync.yml:19
Monday 29 July 2024 08:18:18 -0400 (0:00:00.017) 0:00:23.646 ***********
ok: [managed_node1] => {
"changed": false,
"content": "dG90ZW0gewogICAgdmVyc2lvbjogMgogICAgY2x1c3Rlcl9uYW1lOiB0ZXN0LWNsdXN0ZXIKICAgIHRyYW5zcG9ydDoga25ldAogICAgY3J5cHRvX2NpcGhlcjogYWVzMjU2CiAgICBjcnlwdG9faGFzaDogc2hhMjU2Cn0KCm5vZGVsaXN0IHsKICAgIG5vZGUgewogICAgICAgIHJpbmcwX2FkZHI6IGxvY2FsaG9zdAogICAgICAgIG5hbWU6IGxvY2FsaG9zdAogICAgICAgIG5vZGVpZDogMQogICAgfQp9CgpxdW9ydW0gewogICAgcHJvdmlkZXI6IGNvcm9zeW5jX3ZvdGVxdW9ydW0KfQoKbG9nZ2luZyB7CiAgICB0b19sb2dmaWxlOiB5ZXMKICAgIGxvZ2ZpbGU6IC92YXIvbG9nL2NsdXN0ZXIvY29yb3N5bmMubG9nCiAgICB0b19zeXNsb2c6IHllcwogICAgdGltZXN0YW1wOiBvbgp9Cg==",
"encoding": "base64",
"source": "/tmp/ansible.eyt1mp6k_ha_cluster_corosync_conf"
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute corosync.conf file] ****
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-corosync.yml:26
Monday 29 July 2024 08:18:18 -0400 (0:00:00.344) 0:00:23.991 ***********
ok: [managed_node1] => {
"changed": false,
"checksum": "05d2ec2a2bfa233bb2b4ace4aae02b42cafc012b",
"dest": "/etc/corosync/corosync.conf",
"gid": 0,
"group": "root",
"mode": "0644",
"owner": "root",
"path": "/etc/corosync/corosync.conf",
"secontext": "system_u:object_r:etc_t:s0",
"size": 388,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Remove a corosync.conf tempfile] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-corosync.yml:36
Monday 29 July 2024 08:18:19 -0400 (0:00:00.597) 0:00:24.588 ***********
changed: [managed_node1] => {
"changed": true,
"path": "/tmp/ansible.eyt1mp6k_ha_cluster_corosync_conf",
"state": "absent"
}
TASK [fedora.linux_system_roles.ha_cluster : Cluster auth] *********************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:61
Monday 29 July 2024 08:18:19 -0400 (0:00:00.356) 0:00:24.945 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-auth.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Check pcs auth status] ************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-auth.yml:3
Monday 29 July 2024 08:18:19 -0400 (0:00:00.043) 0:00:24.988 ***********
ok: [managed_node1] => {
"changed": false,
"cmd": [
"pcs",
"status",
"pcsd",
"--",
"localhost"
],
"delta": "0:00:00.527669",
"end": "2024-07-29 08:18:20.232520",
"failed_when_result": false,
"rc": 0,
"start": "2024-07-29 08:18:19.704851"
}
STDOUT:
localhost: Online
TASK [fedora.linux_system_roles.ha_cluster : Run pcs auth] *********************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-auth.yml:23
Monday 29 July 2024 08:18:20 -0400 (0:00:00.886) 0:00:25.875 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute cluster shared keys] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:66
Monday 29 July 2024 08:18:20 -0400 (0:00:00.020) 0:00:25.895 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Get corosync authkey] *************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:3
Monday 29 July 2024 08:18:20 -0400 (0:00:00.048) 0:00:25.944 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Check if key exists on the controller corosync authkey] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:16
Monday 29 July 2024 08:18:20 -0400 (0:00:00.039) 0:00:25.984 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Slurp existing key from the controller corosync authkey] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:21
Monday 29 July 2024 08:18:20 -0400 (0:00:00.023) 0:00:26.008 ***********
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Use the slurped key from the controller corosync authkey] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:29
Monday 29 July 2024 08:18:20 -0400 (0:00:00.026) 0:00:26.034 ***********
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Generate key using OpenSSL corosync authkey] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:44
Monday 29 July 2024 08:18:20 -0400 (0:00:00.021) 0:00:26.056 ***********
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch generated corosync authkey] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:50
Monday 29 July 2024 08:18:20 -0400 (0:00:00.355) 0:00:26.411 ***********
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Check if key exists on cluster nodes corosync authkey] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:60
Monday 29 July 2024 08:18:20 -0400 (0:00:00.027) 0:00:26.438 ***********
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1722255278.9013352,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 8,
"charset": "binary",
"checksum": "2af468ad87ee36eda6dcb9a31c511d244cb700e7",
"ctime": 1722255269.5323455,
"dev": 51713,
"device_type": 0,
"executable": false,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 113246440,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/octet-stream",
"mode": "0400",
"mtime": 1722255269.2653458,
"nlink": 1,
"path": "/etc/corosync/authkey",
"pw_name": "root",
"readable": true,
"rgrp": false,
"roth": false,
"rusr": true,
"size": 256,
"uid": 0,
"version": "590164419",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": false,
"xgrp": false,
"xoth": false,
"xusr": false
}
}
TASK [fedora.linux_system_roles.ha_cluster : Slurp existing key from cluster nodes corosync authkey] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:65
Monday 29 July 2024 08:18:21 -0400 (0:00:00.362) 0:00:26.801 ***********
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Use the slurped key from cluster nodes corosync authkey] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:73
Monday 29 July 2024 08:18:21 -0400 (0:00:00.357) 0:00:27.159 ***********
ok: [managed_node1 -> managed_node1] => (item=None) => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
ok: [managed_node1 -> {{ item }}] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute corosync authkey] ******
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:11
Monday 29 July 2024 08:18:21 -0400 (0:00:00.035) 0:00:27.194 ***********
ok: [managed_node1] => {
"changed": false,
"checksum": "2af468ad87ee36eda6dcb9a31c511d244cb700e7",
"dest": "/etc/corosync/authkey",
"gid": 0,
"group": "root",
"mode": "0400",
"owner": "root",
"path": "/etc/corosync/authkey",
"secontext": "system_u:object_r:etc_t:s0",
"size": 256,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Get pacemaker authkey] ************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:20
Monday 29 July 2024 08:18:22 -0400 (0:00:00.596) 0:00:27.791 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Check if key exists on the controller pacemaker authkey] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:16
Monday 29 July 2024 08:18:22 -0400 (0:00:00.041) 0:00:27.833 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Slurp existing key from the controller pacemaker authkey] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:21
Monday 29 July 2024 08:18:22 -0400 (0:00:00.022) 0:00:27.855 ***********
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Use the slurped key from the controller pacemaker authkey] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:29
Monday 29 July 2024 08:18:22 -0400 (0:00:00.021) 0:00:27.877 ***********
skipping: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Generate key using OpenSSL pacemaker authkey] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:44
Monday 29 July 2024 08:18:22 -0400 (0:00:00.022) 0:00:27.899 ***********
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch generated pacemaker authkey] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:50
Monday 29 July 2024 08:18:22 -0400 (0:00:00.353) 0:00:28.253 ***********
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Check if key exists on cluster nodes pacemaker authkey] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:60
Monday 29 July 2024 08:18:22 -0400 (0:00:00.026) 0:00:28.279 ***********
ok: [managed_node1] => {
"changed": false,
"stat": {
"atime": 1722255375.6231945,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 8,
"charset": "binary",
"checksum": "85460fa6e595156e5797dd33fd7c679e23eb67bb",
"ctime": 1722255271.2453437,
"dev": 51713,
"device_type": 0,
"executable": false,
"exists": true,
"gid": 189,
"gr_name": "haclient",
"inode": 127926484,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/octet-stream",
"mode": "0400",
"mtime": 1722255270.984344,
"nlink": 1,
"path": "/etc/pacemaker/authkey",
"pw_name": "hacluster",
"readable": true,
"rgrp": false,
"roth": false,
"rusr": true,
"size": 256,
"uid": 189,
"version": "4017451340",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": false,
"xgrp": false,
"xoth": false,
"xusr": false
}
}
TASK [fedora.linux_system_roles.ha_cluster : Slurp existing key from cluster nodes pacemaker authkey] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:65
Monday 29 July 2024 08:18:23 -0400 (0:00:00.362) 0:00:28.642 ***********
ok: [managed_node1] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Use the slurped key from cluster nodes pacemaker authkey] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:73
Monday 29 July 2024 08:18:23 -0400 (0:00:00.379) 0:00:29.022 ***********
ok: [managed_node1 -> managed_node1] => (item=None) => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
ok: [managed_node1 -> {{ item }}] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute pacemaker authkey] *****
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:28
Monday 29 July 2024 08:18:23 -0400 (0:00:00.034) 0:00:29.056 ***********
ok: [managed_node1] => {
"changed": false,
"checksum": "85460fa6e595156e5797dd33fd7c679e23eb67bb",
"dest": "/etc/pacemaker/authkey",
"gid": 189,
"group": "haclient",
"mode": "0400",
"owner": "hacluster",
"path": "/etc/pacemaker/authkey",
"secontext": "system_u:object_r:etc_t:s0",
"size": 256,
"state": "file",
"uid": 189
}
TASK [fedora.linux_system_roles.ha_cluster : Remove qdevice certificates [CLI]] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:37
Monday 29 July 2024 08:18:24 -0400 (0:00:00.597) 0:00:29.654 ***********
changed: [managed_node1] => {
"changed": true,
"cmd": [
"pcs",
"--",
"qdevice",
"net-client",
"destroy"
],
"delta": "0:00:00.496673",
"end": "2024-07-29 08:18:24.859967",
"rc": 0,
"start": "2024-07-29 08:18:24.363294"
}
TASK [fedora.linux_system_roles.ha_cluster : Remove qdevice certificates [API]] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:46
Monday 29 July 2024 08:18:24 -0400 (0:00:00.847) 0:00:30.501 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Obtain and distribute qdevice certificates [CLI]] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:54
Monday 29 July 2024 08:18:24 -0400 (0:00:00.023) 0:00:30.524 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Obtain and distribute qdevice certificates [API]] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:71
Monday 29 July 2024 08:18:24 -0400 (0:00:00.019) 0:00:30.544 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Enable or disable cluster services on boot] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:72
Monday 29 July 2024 08:18:24 -0400 (0:00:00.018) 0:00:30.562 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_common/cluster-enable-disable.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Enable or disable configured cluster services on boot] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_common/cluster-enable-disable.yml:3
Monday 29 July 2024 08:18:25 -0400 (0:00:00.046) 0:00:30.609 ***********
ok: [managed_node1] => (item=corosync) => {
"ansible_loop_var": "item",
"changed": false,
"enabled": true,
"item": "corosync",
"name": "corosync",
"status": {
"ActiveEnterTimestamp": "Mon 2024-07-29 08:14:39 EDT",
"ActiveEnterTimestampMonotonic": "312304834",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "basic.target system.slice network-online.target sysinit.target systemd-journald.socket",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Mon 2024-07-29 08:14:38 EDT",
"AssertTimestampMonotonic": "311900475",
"Before": "shutdown.target multi-user.target pacemaker.service",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Mon 2024-07-29 08:14:38 EDT",
"ConditionTimestampMonotonic": "311900448",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroup": "/system.slice/corosync.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Corosync Cluster Engine",
"DevicePolicy": "auto",
"Documentation": "man:corosync man:corosync.conf man:corosync_overview",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/corosync (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "20987",
"ExecMainStartTimestamp": "Mon 2024-07-29 08:14:38 EDT",
"ExecMainStartTimestampMonotonic": "311901701",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/corosync ; argv[]=/usr/sbin/corosync -f $COROSYNC_OPTIONS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStop": "{ path=/usr/sbin/corosync-cfgtool ; argv[]=/usr/sbin/corosync-cfgtool -H --force ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/corosync.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "corosync.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Mon 2024-07-29 08:14:38 EDT",
"InactiveExitTimestampMonotonic": "311901745",
"InvocationID": "10b8314ce99345da893fd801d3ae13a2",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14004",
"LimitNPROCSoft": "14004",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14004",
"LimitSIGPENDINGSoft": "14004",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "20987",
"MemoryAccounting": "yes",
"MemoryCurrent": "145793024",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "corosync.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"RequiredBy": "pacemaker.service",
"Requires": "system.slice network-online.target sysinit.target",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Mon 2024-07-29 08:14:39 EDT",
"StateChangeTimestampMonotonic": "312304834",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "9",
"TasksMax": "22406",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogTimestamp": "Mon 2024-07-29 08:14:39 EDT",
"WatchdogTimestampMonotonic": "312304815",
"WatchdogUSec": "0"
}
}
skipping: [managed_node1] => (item=corosync-qdevice) => {
"ansible_loop_var": "item",
"changed": false,
"item": "corosync-qdevice",
"skip_reason": "Conditional result was False"
}
ok: [managed_node1] => (item=pacemaker) => {
"ansible_loop_var": "item",
"changed": false,
"enabled": true,
"item": "pacemaker",
"name": "pacemaker",
"status": {
"ActiveEnterTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"ActiveEnterTimestampMonotonic": "313203297",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "rsyslog.service systemd-journald.socket sysinit.target system.slice network.target time-sync.target corosync.service resource-agents-deps.target basic.target dbus.service",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"AssertTimestampMonotonic": "313201827",
"Before": "shutdown.target multi-user.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"ConditionTimestampMonotonic": "313201826",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroup": "/system.slice/pacemaker.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Pacemaker High Availability Cluster Manager",
"DevicePolicy": "auto",
"Documentation": "man:pacemakerd https://clusterlabs.org/pacemaker/doc/",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/sbd (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "21250",
"ExecMainStartTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"ExecMainStartTimestampMonotonic": "313203253",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/pacemakerd ; argv[]=/usr/sbin/pacemakerd ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/pacemaker.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "pacemaker.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"InactiveExitTimestampMonotonic": "313203297",
"InvocationID": "4c3b126d203c456db5b92445c9cbb696",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "process",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14004",
"LimitNPROCSoft": "14004",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14004",
"LimitSIGPENDINGSoft": "14004",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "21250",
"MemoryAccounting": "yes",
"MemoryCurrent": "54763520",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "pacemaker.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "sysinit.target system.slice corosync.service",
"Restart": "on-failure",
"RestartUSec": "1s",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "no",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "25s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"StateChangeTimestampMonotonic": "313203297",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "7",
"TasksMax": "infinity",
"TimeoutStartUSec": "1min",
"TimeoutStopUSec": "30min",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "simple",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"Wants": "dbus.service resource-agents-deps.target",
"WatchdogTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"WatchdogTimestampMonotonic": "313203292",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Get services status - detect SBD] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_common/cluster-enable-disable.yml:16
Monday 29 July 2024 08:18:25 -0400 (0:00:00.956) 0:00:31.565 ***********
ok: [managed_node1] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"corosync-notifyd.service": {
"name": "corosync-notifyd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qnetd.service": {
"name": "corosync-qnetd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync.service": {
"name": "corosync.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crm_mon.service": {
"name": "crm_mon.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "unknown"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-pvscan@.service": {
"name": "lvm2-pvscan@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"multipathd.service": {
"name": "multipathd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"pacemaker.service": {
"name": "pacemaker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"pcsd-ruby.service": {
"name": "pcsd-ruby.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"pcsd.service": {
"name": "pcsd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sbd.service": {
"name": "sbd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"sbd_remote.service": {
"name": "sbd_remote.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "unknown"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Enable or disable SBD] ************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_common/cluster-enable-disable.yml:19
Monday 29 July 2024 08:18:27 -0400 (0:00:01.588) 0:00:33.154 ***********
changed: [managed_node1] => {
"changed": true,
"enabled": true,
"name": "sbd",
"status": {
"ActiveEnterTimestampMonotonic": "0",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "inactive",
"After": "iscsi.service sysinit.target system.slice systemd-modules-load.service systemd-journald.socket basic.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "no",
"AssertTimestampMonotonic": "0",
"Before": "shutdown.target pacemaker.service dlm.service",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "no",
"CanStop": "no",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "no",
"ConditionTimestampMonotonic": "0",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Shared-storage based fencing daemon",
"DevicePolicy": "auto",
"Documentation": "man:sbd(8)",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/sbd (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "0",
"ExecMainStartTimestampMonotonic": "0",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/sbd ; argv[]=/usr/sbin/sbd $SBD_OPTS -p /run/sbd.pid watch ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStop": "{ path=/usr/bin/kill ; argv[]=/usr/bin/kill -TERM $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/sbd.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "sbd.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestampMonotonic": "0",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14004",
"LimitNPROCSoft": "14004",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14004",
"LimitSIGPENDINGSoft": "14004",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "sbd.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PIDFile": "/run/sbd.pid",
"PartOf": "corosync.service",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "yes",
"RefuseManualStop": "yes",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "system.slice sysinit.target",
"Restart": "on-abort",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestampMonotonic": "0",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "22406",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "forking",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "disabled",
"UtmpMode": "init",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Start the cluster and reload corosync.conf] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:75
Monday 29 July 2024 08:18:28 -0400 (0:00:00.659) 0:00:33.814 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Get services status - detect corosync-qdevice] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:21
Monday 29 July 2024 08:18:28 -0400 (0:00:00.050) 0:00:33.864 ***********
ok: [managed_node1] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"corosync-notifyd.service": {
"name": "corosync-notifyd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qnetd.service": {
"name": "corosync-qnetd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync.service": {
"name": "corosync.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crm_mon.service": {
"name": "crm_mon.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "unknown"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-pvscan@.service": {
"name": "lvm2-pvscan@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"multipathd.service": {
"name": "multipathd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"pacemaker.service": {
"name": "pacemaker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"pcsd-ruby.service": {
"name": "pcsd-ruby.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"pcsd.service": {
"name": "pcsd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sbd.service": {
"name": "sbd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"sbd_remote.service": {
"name": "sbd_remote.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "unknown"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Stop cluster daemons to reload configuration] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:24
Monday 29 July 2024 08:18:29 -0400 (0:00:01.618) 0:00:35.482 ***********
changed: [managed_node1] => (item=pacemaker) => {
"ansible_loop_var": "item",
"changed": true,
"item": "pacemaker",
"name": "pacemaker",
"state": "stopped",
"status": {
"ActiveEnterTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"ActiveEnterTimestampMonotonic": "313203297",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "sysinit.target dbus.service network.target systemd-journald.socket sbd.service rsyslog.service time-sync.target resource-agents-deps.target system.slice corosync.service basic.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"AssertTimestampMonotonic": "313201827",
"Before": "multi-user.target shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"ConditionTimestampMonotonic": "313201826",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroup": "/system.slice/pacemaker.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Pacemaker High Availability Cluster Manager",
"DevicePolicy": "auto",
"Documentation": "man:pacemakerd https://clusterlabs.org/pacemaker/doc/",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/sbd (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "21250",
"ExecMainStartTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"ExecMainStartTimestampMonotonic": "313203253",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/pacemakerd ; argv[]=/usr/sbin/pacemakerd ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/pacemaker.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "pacemaker.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"InactiveExitTimestampMonotonic": "313203297",
"InvocationID": "4c3b126d203c456db5b92445c9cbb696",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "process",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14004",
"LimitNPROCSoft": "14004",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14004",
"LimitSIGPENDINGSoft": "14004",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "21250",
"MemoryAccounting": "yes",
"MemoryCurrent": "54759424",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "pacemaker.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "sysinit.target sbd.service corosync.service system.slice",
"Restart": "on-failure",
"RestartUSec": "1s",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "no",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "25s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"StateChangeTimestampMonotonic": "313203297",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "7",
"TasksMax": "infinity",
"TimeoutStartUSec": "1min",
"TimeoutStopUSec": "30min",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "simple",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"Wants": "dbus.service resource-agents-deps.target",
"WatchdogTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"WatchdogTimestampMonotonic": "313203292",
"WatchdogUSec": "0"
}
}
skipping: [managed_node1] => (item=corosync-qdevice) => {
"ansible_loop_var": "item",
"changed": false,
"item": "corosync-qdevice",
"skip_reason": "Conditional result was False"
}
changed: [managed_node1] => (item=corosync) => {
"ansible_loop_var": "item",
"changed": true,
"item": "corosync",
"name": "corosync",
"state": "stopped",
"status": {
"ActiveEnterTimestamp": "Mon 2024-07-29 08:14:39 EDT",
"ActiveEnterTimestampMonotonic": "312304834",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "sysinit.target system.slice network-online.target systemd-journald.socket basic.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Mon 2024-07-29 08:14:38 EDT",
"AssertTimestampMonotonic": "311900475",
"Before": "multi-user.target pacemaker.service shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Mon 2024-07-29 08:14:38 EDT",
"ConditionTimestampMonotonic": "311900448",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ConsistsOf": "sbd.service",
"ControlGroup": "/system.slice/corosync.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Corosync Cluster Engine",
"DevicePolicy": "auto",
"Documentation": "man:corosync man:corosync.conf man:corosync_overview",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/corosync (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "20987",
"ExecMainStartTimestamp": "Mon 2024-07-29 08:14:38 EDT",
"ExecMainStartTimestampMonotonic": "311901701",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/corosync ; argv[]=/usr/sbin/corosync -f $COROSYNC_OPTIONS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStop": "{ path=/usr/sbin/corosync-cfgtool ; argv[]=/usr/sbin/corosync-cfgtool -H --force ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/corosync.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "corosync.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Mon 2024-07-29 08:14:38 EDT",
"InactiveExitTimestampMonotonic": "311901745",
"InvocationID": "10b8314ce99345da893fd801d3ae13a2",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14004",
"LimitNPROCSoft": "14004",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14004",
"LimitSIGPENDINGSoft": "14004",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "20987",
"MemoryAccounting": "yes",
"MemoryCurrent": "120594432",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "corosync.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"RequiredBy": "pacemaker.service",
"Requires": "sbd.service sysinit.target system.slice network-online.target",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Mon 2024-07-29 08:14:39 EDT",
"StateChangeTimestampMonotonic": "312304834",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "9",
"TasksMax": "22406",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogTimestamp": "Mon 2024-07-29 08:14:39 EDT",
"WatchdogTimestampMonotonic": "312304815",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Start corosync] *******************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:47
Monday 29 July 2024 08:18:31 -0400 (0:00:01.851) 0:00:37.334 ***********
changed: [managed_node1] => {
"changed": true,
"name": "corosync",
"state": "started",
"status": {
"ActiveEnterTimestamp": "Mon 2024-07-29 08:14:39 EDT",
"ActiveEnterTimestampMonotonic": "312304834",
"ActiveExitTimestamp": "Mon 2024-07-29 08:18:30 EDT",
"ActiveExitTimestampMonotonic": "543872297",
"ActiveState": "inactive",
"After": "sysinit.target system.slice network-online.target systemd-journald.socket basic.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Mon 2024-07-29 08:14:38 EDT",
"AssertTimestampMonotonic": "311900475",
"Before": "multi-user.target pacemaker.service shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Mon 2024-07-29 08:14:38 EDT",
"ConditionTimestampMonotonic": "311900448",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ConsistsOf": "sbd.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Corosync Cluster Engine",
"DevicePolicy": "auto",
"Documentation": "man:corosync man:corosync.conf man:corosync_overview",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/corosync (ignore_errors=yes)",
"ExecMainCode": "1",
"ExecMainExitTimestamp": "Mon 2024-07-29 08:18:31 EDT",
"ExecMainExitTimestampMonotonic": "544689060",
"ExecMainPID": "20987",
"ExecMainStartTimestamp": "Mon 2024-07-29 08:14:38 EDT",
"ExecMainStartTimestampMonotonic": "311901701",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/corosync ; argv[]=/usr/sbin/corosync -f $COROSYNC_OPTIONS ; ignore_errors=no ; start_time=[Mon 2024-07-29 08:14:38 EDT] ; stop_time=[Mon 2024-07-29 08:18:31 EDT] ; pid=20987 ; code=exited ; status=0 }",
"ExecStop": "{ path=/usr/sbin/corosync-cfgtool ; argv[]=/usr/sbin/corosync-cfgtool -H --force ; ignore_errors=no ; start_time=[Mon 2024-07-29 08:18:30 EDT] ; stop_time=[Mon 2024-07-29 08:18:30 EDT] ; pid=51116 ; code=exited ; status=0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/corosync.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "corosync.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestamp": "Mon 2024-07-29 08:18:31 EDT",
"InactiveEnterTimestampMonotonic": "544689192",
"InactiveExitTimestamp": "Mon 2024-07-29 08:14:38 EDT",
"InactiveExitTimestampMonotonic": "311901745",
"InvocationID": "10b8314ce99345da893fd801d3ae13a2",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14004",
"LimitNPROCSoft": "14004",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14004",
"LimitSIGPENDINGSoft": "14004",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "corosync.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"RequiredBy": "pacemaker.service",
"Requires": "sbd.service sysinit.target system.slice network-online.target",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Mon 2024-07-29 08:18:31 EDT",
"StateChangeTimestampMonotonic": "544689192",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "22406",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Reload corosync configuration] ****
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:56
Monday 29 July 2024 08:18:32 -0400 (0:00:00.890) 0:00:38.224 ***********
ok: [managed_node1] => {
"changed": false,
"cmd": [
"corosync-cfgtool",
"-R"
],
"delta": "0:00:00.011906",
"end": "2024-07-29 08:18:32.941679",
"rc": 0,
"start": "2024-07-29 08:18:32.929773"
}
STDOUT:
Reloading corosync.conf...
Done
TASK [fedora.linux_system_roles.ha_cluster : Start corosync-qdevice] ***********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:63
Monday 29 July 2024 08:18:32 -0400 (0:00:00.360) 0:00:38.585 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Start pacemaker] ******************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:69
Monday 29 July 2024 08:18:33 -0400 (0:00:00.018) 0:00:38.603 ***********
changed: [managed_node1] => {
"changed": true,
"name": "pacemaker",
"state": "started",
"status": {
"ActiveEnterTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"ActiveEnterTimestampMonotonic": "313203297",
"ActiveExitTimestamp": "Mon 2024-07-29 08:18:30 EDT",
"ActiveExitTimestampMonotonic": "543332388",
"ActiveState": "inactive",
"After": "sysinit.target dbus.service network.target systemd-journald.socket sbd.service rsyslog.service time-sync.target resource-agents-deps.target system.slice corosync.service basic.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"AssertTimestampMonotonic": "313201827",
"Before": "multi-user.target shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"ConditionTimestampMonotonic": "313201826",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Pacemaker High Availability Cluster Manager",
"DevicePolicy": "auto",
"Documentation": "man:pacemakerd https://clusterlabs.org/pacemaker/doc/",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/sbd (ignore_errors=yes)",
"ExecMainCode": "1",
"ExecMainExitTimestamp": "Mon 2024-07-29 08:18:30 EDT",
"ExecMainExitTimestampMonotonic": "543383136",
"ExecMainPID": "21250",
"ExecMainStartTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"ExecMainStartTimestampMonotonic": "313203253",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/pacemakerd ; argv[]=/usr/sbin/pacemakerd ; ignore_errors=no ; start_time=[Mon 2024-07-29 08:14:40 EDT] ; stop_time=[Mon 2024-07-29 08:18:30 EDT] ; pid=21250 ; code=exited ; status=0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/pacemaker.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "pacemaker.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestamp": "Mon 2024-07-29 08:18:30 EDT",
"InactiveEnterTimestampMonotonic": "543383208",
"InactiveExitTimestamp": "Mon 2024-07-29 08:14:40 EDT",
"InactiveExitTimestampMonotonic": "313203297",
"InvocationID": "4c3b126d203c456db5b92445c9cbb696",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "process",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14004",
"LimitNPROCSoft": "14004",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14004",
"LimitSIGPENDINGSoft": "14004",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "pacemaker.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "sysinit.target sbd.service corosync.service system.slice",
"Restart": "on-failure",
"RestartUSec": "1s",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "no",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "25s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Mon 2024-07-29 08:18:30 EDT",
"StateChangeTimestampMonotonic": "543383208",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "infinity",
"TimeoutStartUSec": "1min",
"TimeoutStopUSec": "30min",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "simple",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"Wants": "dbus.service resource-agents-deps.target",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Wait for the cluster to fully start and form membership] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:75
Monday 29 July 2024 08:18:33 -0400 (0:00:00.518) 0:00:39.122 ***********
ok: [managed_node1] => {
"changed": false,
"cmd": [
"pcs",
"cluster",
"start",
"--all",
"--wait"
],
"delta": "0:00:26.346826",
"end": "2024-07-29 08:19:00.184736",
"rc": 0,
"start": "2024-07-29 08:18:33.837910"
}
STDOUT:
localhost: Starting Cluster...
Waiting for node(s) to start...
localhost: Started
TASK [fedora.linux_system_roles.ha_cluster : List pacemaker nodes] *************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:89
Monday 29 July 2024 08:19:00 -0400 (0:00:26.706) 0:01:05.828 ***********
ok: [managed_node1] => {
"changed": false,
"cmd": "set -euo pipefail; crm_mon -X | xmllint --xpath '/crm_mon/nodes/node/@name' - | sed -E 's/\\s*name=\"([^\"]+)\"\\s*/\\1\\n/g'\n",
"delta": "0:00:00.021103",
"end": "2024-07-29 08:19:00.562071",
"rc": 0,
"start": "2024-07-29 08:19:00.540968"
}
STDOUT:
localhost
TASK [fedora.linux_system_roles.ha_cluster : Purge removed nodes from pacemaker's cache] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:111
Monday 29 July 2024 08:19:00 -0400 (0:00:00.373) 0:01:06.202 ***********
skipping: [managed_node1] => (item=localhost) => {
"ansible_loop_var": "item",
"changed": false,
"item": "localhost",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Create and push CIB] **************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:78
Monday 29 July 2024 08:19:00 -0400 (0:00:00.027) 0:01:06.230 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Create a tempfile for original CIB] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:10
Monday 29 July 2024 08:19:00 -0400 (0:00:00.072) 0:01:06.302 ***********
changed: [managed_node1] => {
"changed": true,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.jgvggnjr_ha_cluster_original_cib_xml",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Create a tempfile for new CIB] ****
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:18
Monday 29 July 2024 08:19:01 -0400 (0:00:00.365) 0:01:06.667 ***********
changed: [managed_node1] => {
"changed": true,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.okf8wbec_ha_cluster_cib_xml",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch CIB configuration] **********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:26
Monday 29 July 2024 08:19:01 -0400 (0:00:00.367) 0:01:07.034 ***********
ok: [managed_node1] => {
"changed": false,
"cmd": [
"cibadmin",
"--query"
],
"delta": "0:00:00.012151",
"end": "2024-07-29 08:19:01.763855",
"rc": 0,
"start": "2024-07-29 08:19:01.751704"
}
STDOUT:
TASK [fedora.linux_system_roles.ha_cluster : Write CIB configuration] **********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:33
Monday 29 July 2024 08:19:01 -0400 (0:00:00.376) 0:01:07.411 ***********
changed: [managed_node1] => (item=/tmp/ansible.okf8wbec_ha_cluster_cib_xml) => {
"ansible_loop_var": "item",
"changed": true,
"checksum": "c4c2a3ea7de68d81b499c1a8af4cec37b44aa47f",
"dest": "/tmp/ansible.okf8wbec_ha_cluster_cib_xml",
"gid": 0,
"group": "root",
"item": "/tmp/ansible.okf8wbec_ha_cluster_cib_xml",
"md5sum": "395bb6b7137e697f609c38897ed1d6d7",
"mode": "0600",
"owner": "root",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 17372,
"src": "/root/.ansible/tmp/ansible-tmp-1722255541.8746302-12183-229687199407639/source",
"state": "file",
"uid": 0
}
changed: [managed_node1] => (item=/tmp/ansible.jgvggnjr_ha_cluster_original_cib_xml) => {
"ansible_loop_var": "item",
"changed": true,
"checksum": "c4c2a3ea7de68d81b499c1a8af4cec37b44aa47f",
"dest": "/tmp/ansible.jgvggnjr_ha_cluster_original_cib_xml",
"gid": 0,
"group": "root",
"item": "/tmp/ansible.jgvggnjr_ha_cluster_original_cib_xml",
"md5sum": "395bb6b7137e697f609c38897ed1d6d7",
"mode": "0600",
"owner": "root",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 17372,
"src": "/root/.ansible/tmp/ansible-tmp-1722255542.5584123-12183-40595517960488/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Purge new CIB configuration] ******
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:51
Monday 29 July 2024 08:19:03 -0400 (0:00:01.388) 0:01:08.799 ***********
changed: [managed_node1] => {
"changed": true,
"cmd": [
"cibadmin",
"--force",
"--delete-all",
"--xpath",
"/cib/configuration/*[not(\n self::crm_config or\n self::nodes or\n self::resources or\n self::constraints\n)] | /cib/configuration/*[self::resources or self::constraints]/* | /cib/configuration/nodes/*/* | /cib/configuration/crm_config//nvpair[not(\n @name=\"cluster-infrastructure\" or\n @name=\"cluster-name\" or\n @name=\"dc-version\" or\n @name=\"have-watchdog\" or\n @name=\"last-lrm-refresh\" or\n @name=\"stonith-watchdog-timeout\"\n)]"
],
"delta": "0:00:00.028611",
"end": "2024-07-29 08:19:03.548614",
"rc": 0,
"start": "2024-07-29 08:19:03.520003"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure cluster properties] *****
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:85
Monday 29 July 2024 08:19:03 -0400 (0:00:00.393) 0:01:09.193 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure node attributes] ********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:95
Monday 29 July 2024 08:19:03 -0400 (0:00:00.020) 0:01:09.213 ***********
TASK [fedora.linux_system_roles.ha_cluster : Configure node utilization] *******
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:102
Monday 29 July 2024 08:19:03 -0400 (0:00:00.019) 0:01:09.233 ***********
TASK [fedora.linux_system_roles.ha_cluster : Configure resource defaults] ******
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:115
Monday 29 July 2024 08:19:03 -0400 (0:00:00.017) 0:01:09.251 ***********
TASK [fedora.linux_system_roles.ha_cluster : Configure resource operation defaults] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:127
Monday 29 July 2024 08:19:03 -0400 (0:00:00.017) 0:01:09.268 ***********
TASK [fedora.linux_system_roles.ha_cluster : Extract primitive to bundle mapping] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:137
Monday 29 July 2024 08:19:03 -0400 (0:00:00.017) 0:01:09.285 ***********
ok: [managed_node1] => {
"ansible_facts": {
"__ha_cluster_primitive_bundle_map": {}
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Configure cluster bundle resources] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:141
Monday 29 July 2024 08:19:03 -0400 (0:00:00.025) 0:01:09.311 ***********
TASK [fedora.linux_system_roles.ha_cluster : Configure cluster resources] ******
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:147
Monday 29 July 2024 08:19:03 -0400 (0:00:00.017) 0:01:09.328 ***********
TASK [fedora.linux_system_roles.ha_cluster : Configure cluster resource groups] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:154
Monday 29 July 2024 08:19:03 -0400 (0:00:00.017) 0:01:09.345 ***********
TASK [fedora.linux_system_roles.ha_cluster : Configure cluster resource clones] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:160
Monday 29 July 2024 08:19:03 -0400 (0:00:00.018) 0:01:09.363 ***********
TASK [fedora.linux_system_roles.ha_cluster : Configure stonith levels] *********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:167
Monday 29 July 2024 08:19:03 -0400 (0:00:00.018) 0:01:09.382 ***********
TASK [fedora.linux_system_roles.ha_cluster : Configure resource location constraints] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:175
Monday 29 July 2024 08:19:03 -0400 (0:00:00.017) 0:01:09.399 ***********
TASK [fedora.linux_system_roles.ha_cluster : Configure resource colocation constraints] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:182
Monday 29 July 2024 08:19:03 -0400 (0:00:00.018) 0:01:09.417 ***********
TASK [fedora.linux_system_roles.ha_cluster : Configure resource set colocation constraints] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:190
Monday 29 July 2024 08:19:03 -0400 (0:00:00.016) 0:01:09.434 ***********
TASK [fedora.linux_system_roles.ha_cluster : Configure resource order constraints] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:200
Monday 29 July 2024 08:19:03 -0400 (0:00:00.016) 0:01:09.451 ***********
TASK [fedora.linux_system_roles.ha_cluster : Configure resource set order constraints] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:208
Monday 29 July 2024 08:19:03 -0400 (0:00:00.019) 0:01:09.470 ***********
TASK [fedora.linux_system_roles.ha_cluster : Configure resource ticket constraints] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:218
Monday 29 July 2024 08:19:03 -0400 (0:00:00.017) 0:01:09.488 ***********
TASK [fedora.linux_system_roles.ha_cluster : Configure resource set ticket constraints] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:226
Monday 29 July 2024 08:19:03 -0400 (0:00:00.016) 0:01:09.505 ***********
TASK [fedora.linux_system_roles.ha_cluster : Configure acls] *******************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:236
Monday 29 July 2024 08:19:03 -0400 (0:00:00.017) 0:01:09.522 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cib-acls.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Configure ACL roles] **************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cib-acls.yml:3
Monday 29 July 2024 08:19:03 -0400 (0:00:00.038) 0:01:09.560 ***********
TASK [fedora.linux_system_roles.ha_cluster : Configure ACL users] **************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cib-acls.yml:28
Monday 29 July 2024 08:19:03 -0400 (0:00:00.019) 0:01:09.580 ***********
TASK [fedora.linux_system_roles.ha_cluster : Configure ACL groups] *************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cib-acls.yml:44
Monday 29 July 2024 08:19:04 -0400 (0:00:00.020) 0:01:09.600 ***********
TASK [fedora.linux_system_roles.ha_cluster : Create a tempfile for CIB diff] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:243
Monday 29 July 2024 08:19:04 -0400 (0:00:00.018) 0:01:09.618 ***********
changed: [managed_node1] => {
"changed": true,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.okyagi28_ha_cluster_cib_diff",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Compare new and original CIB] *****
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:251
Monday 29 July 2024 08:19:04 -0400 (0:00:00.370) 0:01:09.989 ***********
changed: [managed_node1] => {
"changed": true,
"cmd": [
"crm_diff",
"--no-version",
"--original",
"/tmp/ansible.jgvggnjr_ha_cluster_original_cib_xml",
"--new",
"/tmp/ansible.okf8wbec_ha_cluster_cib_xml"
],
"delta": "0:00:00.007128",
"end": "2024-07-29 08:19:04.717893",
"failed_when_result": false,
"rc": 1,
"start": "2024-07-29 08:19:04.710765"
}
STDOUT:
MSG:
non-zero return code
TASK [fedora.linux_system_roles.ha_cluster : Write CIB diff to its tempfile] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:264
Monday 29 July 2024 08:19:04 -0400 (0:00:00.378) 0:01:10.367 ***********
changed: [managed_node1] => {
"changed": true,
"checksum": "3e1c9ebcf787f7b75eaed83a8c33021580e3eea7",
"dest": "/tmp/ansible.okyagi28_ha_cluster_cib_diff",
"gid": 0,
"group": "root",
"md5sum": "4fea5044c9c2853edeb2a78b166ee6be",
"mode": "0600",
"owner": "root",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 5669,
"src": "/root/.ansible/tmp/ansible-tmp-1722255544.8290594-12284-133134522357089/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Push CIB diff to the cluster if it has any changes] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:275
Monday 29 July 2024 08:19:05 -0400 (0:00:00.700) 0:01:11.068 ***********
changed: [managed_node1] => {
"changed": true,
"cmd": [
"cibadmin",
"--verbose",
"--patch",
"--xml-file",
"/tmp/ansible.okyagi28_ha_cluster_cib_diff"
],
"delta": "0:00:00.021295",
"end": "2024-07-29 08:19:05.805380",
"rc": 0,
"start": "2024-07-29 08:19:05.784085"
}
TASK [fedora.linux_system_roles.ha_cluster : Remove CIB tempfiles] *************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:284
Monday 29 July 2024 08:19:05 -0400 (0:00:00.387) 0:01:11.455 ***********
changed: [managed_node1] => (item={'changed': True, 'path': '/tmp/ansible.okf8wbec_ha_cluster_cib_xml', 'uid': 0, 'gid': 0, 'owner': 'root', 'group': 'root', 'mode': '0600', 'state': 'file', 'secontext': 'unconfined_u:object_r:user_tmp_t:s0', 'size': 0, 'failed': False}) => {
"ansible_loop_var": "item",
"changed": true,
"item": {
"changed": true,
"failed": false,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.okf8wbec_ha_cluster_cib_xml",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
},
"path": "/tmp/ansible.okf8wbec_ha_cluster_cib_xml",
"state": "absent"
}
changed: [managed_node1] => (item={'changed': True, 'path': '/tmp/ansible.jgvggnjr_ha_cluster_original_cib_xml', 'uid': 0, 'gid': 0, 'owner': 'root', 'group': 'root', 'mode': '0600', 'state': 'file', 'secontext': 'unconfined_u:object_r:user_tmp_t:s0', 'size': 0, 'failed': False}) => {
"ansible_loop_var": "item",
"changed": true,
"item": {
"changed": true,
"failed": false,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.jgvggnjr_ha_cluster_original_cib_xml",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
},
"path": "/tmp/ansible.jgvggnjr_ha_cluster_original_cib_xml",
"state": "absent"
}
changed: [managed_node1] => (item={'changed': True, 'path': '/tmp/ansible.okyagi28_ha_cluster_cib_diff', 'uid': 0, 'gid': 0, 'owner': 'root', 'group': 'root', 'mode': '0600', 'state': 'file', 'secontext': 'unconfined_u:object_r:user_tmp_t:s0', 'size': 0, 'failed': False}) => {
"ansible_loop_var": "item",
"changed": true,
"item": {
"changed": true,
"failed": false,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.okyagi28_ha_cluster_cib_diff",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
},
"path": "/tmp/ansible.okyagi28_ha_cluster_cib_diff",
"state": "absent"
}
TASK [fedora.linux_system_roles.ha_cluster : Remove cluster configuration] *****
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:87
Monday 29 July 2024 08:19:06 -0400 (0:00:01.051) 0:01:12.507 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Remove fence-virt authkey] ********
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:90
Monday 29 July 2024 08:19:06 -0400 (0:00:00.018) 0:01:12.525 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure qnetd] ******************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:95
Monday 29 July 2024 08:19:06 -0400 (0:00:00.018) 0:01:12.544 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml for managed_node1
TASK [fedora.linux_system_roles.ha_cluster : Remove qnetd configuration] *******
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml:3
Monday 29 July 2024 08:19:06 -0400 (0:00:00.027) 0:01:12.571 ***********
changed: [managed_node1] => {
"changed": true,
"cmd": [
"pcs",
"--force",
"--",
"qdevice",
"destroy",
"net"
],
"delta": "0:00:01.018813",
"end": "2024-07-29 08:19:08.301214",
"rc": 0,
"start": "2024-07-29 08:19:07.282401"
}
STDOUT:
Stopping quorum device...
quorum device stopped
quorum device disabled
Quorum device 'net' configuration files removed
TASK [fedora.linux_system_roles.ha_cluster : Setup qnetd] **********************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml:16
Monday 29 July 2024 08:19:08 -0400 (0:00:01.373) 0:01:13.945 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Enable or disable qnetd service on boot] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml:26
Monday 29 July 2024 08:19:08 -0400 (0:00:00.018) 0:01:13.964 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Slurp SBD config file] ***************************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:40
Monday 29 July 2024 08:19:08 -0400 (0:00:00.019) 0:01:13.983 ***********
ok: [managed_node1] => {
"changed": false,
"content": "IwojIEFuc2libGUgbWFuYWdlZAojCiMgc3lzdGVtX3JvbGU6aGFfY2x1c3RlcgoKCgojIyBUeXBlOiBzdHJpbmcKIyMgRGVmYXVsdDogIiIKIwojIFNCRF9ERVZJQ0Ugc3BlY2lmaWVzIHRoZSBkZXZpY2VzIHRvIHVzZSBmb3IgZXhjaGFuZ2luZyBzYmQgbWVzc2FnZXMKIyBhbmQgdG8gbW9uaXRvci4gSWYgc3BlY2lmeWluZyBtb3JlIHRoYW4gb25lIHBhdGgsIHVzZSAiOyIgYXMKIyBzZXBhcmF0b3IuCiMKU0JEX0RFVklDRT0iIgoKIyMgVHlwZTogeWVzbm8KIyMgRGVmYXVsdDogeWVzCiMKIyBXaGV0aGVyIHRvIGVuYWJsZSB0aGUgcGFjZW1ha2VyIGludGVncmF0aW9uLgojClNCRF9QQUNFTUFLRVI9eWVzCgojIyBUeXBlOiBhbHdheXMgLyBjbGVhbgojIyBEZWZhdWx0OiBhbHdheXMKIwojIFNwZWNpZnkgdGhlIHN0YXJ0IG1vZGUgZm9yIHNiZC4gU2V0dGluZyB0aGlzIHRvICJjbGVhbiIgd2lsbCBvbmx5CiMgYWxsb3cgc2JkIHRvIHN0YXJ0IGlmIGl0IHdhcyBub3QgcHJldmlvdXNseSBmZW5jZWQuIFNlZSB0aGUgLVMgb3B0aW9uCiMgaW4gdGhlIG1hbiBwYWdlLgojClNCRF9TVEFSVE1PREU9ImFsd2F5cyIKCiMjIFR5cGU6IHllc25vIC8gaW50ZWdlcgojIyBEZWZhdWx0OiBubwojCiMgV2hldGhlciB0byBkZWxheSBhZnRlciBzdGFydGluZyBzYmQgb24gYm9vdCBmb3IgIm1zZ3dhaXQiIHNlY29uZHMuCiMgVGhpcyBtYXkgYmUgbmVjZXNzYXJ5IGlmIHlvdXIgY2x1c3RlciBub2RlcyByZWJvb3Qgc28gZmFzdCB0aGF0IHRoZQojIG90aGVyIG5vZGVzIGFyZSBzdGlsbCB3YWl0aW5nIGluIHRoZSBmZW5jZSBhY2tub3dsZWRnZW1lbnQgcGhhc2UuCiMgVGhpcyBpcyBhbiBvY2Nhc2lvbmFsIGlzc3VlIHdpdGggdmlydHVhbCBtYWNoaW5lcy4KIwojIFRoaXMgY2FuIGFsc28gYmUgZW5hYmxlZCBieSBiZWluZyBzZXQgdG8gYSBzcGVjaWZpYyBkZWxheSB2YWx1ZSwgaW4KIyBzZWNvbmRzLiBTb21ldGltZXMgYSBsb25nZXIgZGVsYXkgdGhhbiB0aGUgZGVmYXVsdCwgIm1zZ3dhaXQiLCBpcwojIG5lZWRlZCwgZm9yIGV4YW1wbGUgaW4gdGhlIGNhc2VzIHdoZXJlIGl0J3MgY29uc2lkZXJlZCB0byBiZSBzYWZlciB0bwojIHdhaXQgbG9uZ2VyIHRoYW46CiMgY29yb3N5bmMgdG9rZW4gdGltZW91dCArIGNvbnNlbnN1cyB0aW1lb3V0ICsgcGNta19kZWxheV9tYXggKyBtc2d3YWl0CiMKIyBCZSBhd2FyZSB0aGF0IHRoZSBzcGVjaWFsIHZhbHVlICIxIiBtZWFucyAieWVzIiByYXRoZXIgdGhhbiAiMXMiLgojCiMgQ29uc2lkZXIgdGhhdCB5b3UgbWlnaHQgaGF2ZSB0byBhZGFwdCB0aGUgc3RhcnR1cC10aW1lb3V0IGFjY29yZGluZ2x5CiMgaWYgdGhlIGRlZmF1bHQgaXNuJ3Qgc3VmZmljaWVudC4gKFRpbWVvdXRTdGFydFNlYyBmb3Igc3lzdGVtZCkKIwojIFRoaXMgb3B0aW9uIG1heSBiZSBpZ25vcmVkIGF0IGEgbGF0ZXIgcG9pbnQsIG9uY2UgcGFjZW1ha2VyIGhhbmRsZXMKIyB0aGlzIGNhc2UgYmV0dGVyLgojClNCRF9ERUxBWV9TVEFSVD0ibm8iCgojIyBUeXBlOiBzdHJpbmcKIyMgRGVmYXVsdDogL2Rldi93YXRjaGRvZwojCiMgV2F0Y2hkb2cgZGV2aWNlIHRvIHVzZS4gSWYgc2V0IHRvIC9kZXYvbnVsbCwgbm8gd2F0Y2hkb2cgZGV2aWNlIHdpbGwKIyBiZSB1c2VkLgojClNCRF9XQVRDSERPR19ERVY9Ii9kZXYvd2F0Y2hkb2ciCgojIyBUeXBlOiBpbnRlZ2VyCiMjIERlZmF1bHQ6IDUKIwojIEhvdyBsb25nLCBpbiBzZWNvbmRzLCB0aGUgd2F0Y2hkb2cgd2lsbCB3YWl0IGJlZm9yZSBwYW5pY2tpbmcgdGhlCiMgbm9kZSBpZiBuby1vbmUgdGlja2xlcyBpdC4KIwojIFRoaXMgZGVwZW5kcyBtb3N0bHkgb24geW91ciBzdG9yYWdlIGxhdGVuY3k7IHRoZSBtYWpvcml0eSBvZiBkZXZpY2VzCiMgbXVzdCBiZSBzdWNjZXNzZnVsbHkgcmVhZCB3aXRoaW4gdGhpcyB0aW1lLCBvciBlbHNlIHRoZSBub2RlIHdpbGwKIyBzZWxmLWZlbmNlLgojCiMgSWYgeW91ciBzYmQgZGV2aWNlKHMpIHJlc2lkZSBvbiBhIG11bHRpcGF0aCBzZXR1cCBvciBpU0NTSSwgdGhpcwojIHNob3VsZCBiZSB0aGUgdGltZSByZXF1aXJlZCB0byBkZXRlY3QgYSBwYXRoIGZhaWx1cmUuCiMKIyBCZSBhd2FyZSB0aGF0IHdhdGNoZG9nIHRpbWVvdXQgc2V0IGluIHRoZSBvbi1kaXNrIG1ldGFkYXRhIHRha2VzCiMgcHJlY2VkZW5jZS4KIwpTQkRfV0FUQ0hET0dfVElNRU9VVD0iNSIKCiMjIFR5cGU6IHN0cmluZwojIyBEZWZhdWx0OiAiZmx1c2gscmVib290IgojCiMgQWN0aW9ucyB0byBiZSBleGVjdXRlZCB3aGVuIHRoZSB3YXRjaGVycyBkb24ndCB0aW1lbHkgcmVwb3J0IHRvIHRoZSBzYmQKIyBtYWluIHByb2Nlc3Mgb3Igb25lIG9mIHRoZSB3YXRjaGVycyBkZXRlY3RzIHRoYXQgdGhlIG1haW4gcHJvY2VzcyBoYXMKIyBkaWVkLgojCiMgU2V0IHRpbWVvdXQtYWN0aW9uIHRvIGNvbW1hLXNlcGFyYXRlZCBjb21iaW5hdGlvbiBvZgojIG5vZmx1c2h8Zmx1c2ggcGx1cyByZWJvb3R8Y3Jhc2hkdW1wfG9mZi4KIyBJZiBqdXN0IG9uZSBvZiBib3RoIGlzIGdpdmVuIHRoZSBvdGhlciBzdGF5cyBhdCB0aGUgZGVmYXVsdC4KIwojIFRoaXMgZG9lc24ndCBhZmZlY3QgYWN0aW9ucyBsaWtlIG9mZiwgY3Jhc2hkdW1wLCByZWJvb3QgZXhwbGljaXRseQojIHRyaWdnZXJlZCB2aWEgbWVzc2FnZSBzbG90cy4KIyBBbmQgaXQgZG9lcyBhcyB3ZWxsIG5vdCBjb25maWd1cmUgdGhlIGFjdGlvbiBhIHdhdGNoZG9nIHdvdWxkCiMgdHJpZ2dlciBzaG91bGQgaXQgcnVuIG9mZiAodGhlcmUgaXMgbm8gZ2VuZXJpYyBpbnRlcmZhY2UpLgojClNCRF9USU1FT1VUX0FDVElPTj0iZmx1c2gscmVib290IgoKIyMgVHlwZTogeWVzbm8gLyBhdXRvCiMjIERlZmF1bHQ6IGF1dG8KIwojIElmIENQVUFjY291bnRpbmcgaXMgZW5hYmxlZCBkZWZhdWx0IGlzIG5vdCB0byBhc3NpZ24gYW55IFJULWJ1ZGdldAojIHRvIHRoZSBzeXN0ZW0uc2xpY2Ugd2hpY2ggcHJldmVudHMgc2JkIGZyb20gcnVubmluZyBSUi1zY2hlZHVsZWQuCiMKIyBPbmUgd2F5IHRvIGVzY2FwZSB0aGF0IGlzc3VlIGlzIHRvIG1vdmUgc2JkLXByb2Nlc3NlcyBmcm9tIHRoZQojIHNsaWNlIHRoZXkgd2VyZSBvcmlnaW5hbGx5IHN0YXJ0ZWQgdG8gcm9vdC1zbGljZS4KIyBPZiBjb3Vyc2Ugc3RhcnRpbmcgc2JkIGluIGEgY2VydGFpbiBzbGljZSBtaWdodCBiZSBpbnRlbnRpb25hbC4KIyBUaHVzIGluIGF1dG8tbW9kZSBzYmQgd2lsbCBjaGVjayBpZiB0aGUgc2xpY2UgaGFzIFJULWJ1ZGdldCBhc3NpZ25lZC4KIyBJZiB0aGF0IGlzIHRoZSBjYXNlIHNiZCB3aWxsIHN0YXkgaW4gdGhhdCBzbGljZSB3aGlsZSBpdCB3aWxsCiMgYmUgbW92ZWQgdG8gcm9vdC1zbGljZSBvdGhlcndpc2UuCiMKU0JEX01PVkVfVE9fUk9PVF9DR1JPVVA9YXV0bwoKIyMgVHlwZTogeWVzbm8KIyMgRGVmYXVsdDogeWVzCiMKIyBJZiByZXNvdXJjZSBzdGFydHVwIHN5bmNpbmcgaXMgZW5hYmxlZCB0aGVuIHBhY2VtYWtlcmQgaXMKIyBnb25uYSB3YWl0IHRvIGJlIHBpbmdlZCB2aWEgSVBDIGJlZm9yZSBpdCBzdGFydHMgcmVzb3VyY2VzLgojIE9uIHNodXRkb3duIHBhY2VtYWtlcmQgaXMgZ29pbmcgdG8gd2FpdCBpbiBhIHN0YXRlIHdoZXJlIGl0CiMgaGFzIGNsZWFubHkgc2h1dGRvd24gcmVzb3VyY2VzIHRpbGwgc2JkIGZldGNoZXMgdGhhdCBzdGF0ZS4KIwojIFRoZSBkZWZhdWx0IGlzIHNldCB3aGVuIGJ1aWxkaW5nIFNCRCBhbmQgUGFjZW1ha2VyIGZyb20gc291cmNlLgojIEdvaW5nIGZvciAnbm8nIGlzIHNhZmVyIGlmIGl0IGNhbid0IGJlIGFzc3VyZWQgdGhhdCBTQkQgYW5kCiMgUGFjZW1ha2VyIGluc3RhbGxlZCBkbyBib3RoIHN1cHBvcnQgdGhlIHN5bmNocm9uaXphdGlvbiBmZWF0dXJlLgojIFdoZW4gZ29pbmcgd2l0aCAneWVzJyAtIGFsc28gdXNpbmcgcGFja2FnZSBkZXBlbmRlbmNpZXMgdG8KIyBhc3N1cmUgU0JEICYgUGFjZW1ha2VyIGJvdGggc3VwcG9ydCB0aGUgc3luY2hyb25pemF0aW9uCiMgZmVhdHVyZSBhbmQgYXJlIGFzc3VtaW5nIHRoZSBzYW1lIGRlZmF1bHQgLSBhbiBTQkQgY29uZmlndXJhdGlvbgojIGluaGVyaXRlZCB2aWEgYW4gdXBncmFkZSBkb2Vzbid0IGhhdmUgdG8gYmUgYWx0ZXJlZCB0byBzdGlsbAojIGJlbmVmaXQgZnJvbSB0aGUgbmV3IGZlYXR1cmUuCiMKU0JEX1NZTkNfUkVTT1VSQ0VfU1RBUlRVUD15ZXMKCiMjIFR5cGU6IHN0cmluZwojIyBEZWZhdWx0OiAiIgojCiMgQWRkaXRpb25hbCBvcHRpb25zIGZvciBzdGFydGluZyBzYmQKIwpTQkRfT1BUUz0iLW4gbG9jYWxob3N0Igo=",
"encoding": "base64",
"source": "/etc/sysconfig/sbd"
}
TASK [Decode SBD config] *******************************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:45
Monday 29 July 2024 08:19:08 -0400 (0:00:00.350) 0:01:14.333 ***********
ok: [managed_node1] => {
"ansible_facts": {
"__test_sbd_config_lines": [
"#",
"# Ansible managed",
"#",
"# system_role:ha_cluster",
"",
"",
"",
"## Type: string",
"## Default: \"\"",
"#",
"# SBD_DEVICE specifies the devices to use for exchanging sbd messages",
"# and to monitor. If specifying more than one path, use \";\" as",
"# separator.",
"#",
"SBD_DEVICE=\"\"",
"",
"## Type: yesno",
"## Default: yes",
"#",
"# Whether to enable the pacemaker integration.",
"#",
"SBD_PACEMAKER=yes",
"",
"## Type: always / clean",
"## Default: always",
"#",
"# Specify the start mode for sbd. Setting this to \"clean\" will only",
"# allow sbd to start if it was not previously fenced. See the -S option",
"# in the man page.",
"#",
"SBD_STARTMODE=\"always\"",
"",
"## Type: yesno / integer",
"## Default: no",
"#",
"# Whether to delay after starting sbd on boot for \"msgwait\" seconds.",
"# This may be necessary if your cluster nodes reboot so fast that the",
"# other nodes are still waiting in the fence acknowledgement phase.",
"# This is an occasional issue with virtual machines.",
"#",
"# This can also be enabled by being set to a specific delay value, in",
"# seconds. Sometimes a longer delay than the default, \"msgwait\", is",
"# needed, for example in the cases where it's considered to be safer to",
"# wait longer than:",
"# corosync token timeout + consensus timeout + pcmk_delay_max + msgwait",
"#",
"# Be aware that the special value \"1\" means \"yes\" rather than \"1s\".",
"#",
"# Consider that you might have to adapt the startup-timeout accordingly",
"# if the default isn't sufficient. (TimeoutStartSec for systemd)",
"#",
"# This option may be ignored at a later point, once pacemaker handles",
"# this case better.",
"#",
"SBD_DELAY_START=\"no\"",
"",
"## Type: string",
"## Default: /dev/watchdog",
"#",
"# Watchdog device to use. If set to /dev/null, no watchdog device will",
"# be used.",
"#",
"SBD_WATCHDOG_DEV=\"/dev/watchdog\"",
"",
"## Type: integer",
"## Default: 5",
"#",
"# How long, in seconds, the watchdog will wait before panicking the",
"# node if no-one tickles it.",
"#",
"# This depends mostly on your storage latency; the majority of devices",
"# must be successfully read within this time, or else the node will",
"# self-fence.",
"#",
"# If your sbd device(s) reside on a multipath setup or iSCSI, this",
"# should be the time required to detect a path failure.",
"#",
"# Be aware that watchdog timeout set in the on-disk metadata takes",
"# precedence.",
"#",
"SBD_WATCHDOG_TIMEOUT=\"5\"",
"",
"## Type: string",
"## Default: \"flush,reboot\"",
"#",
"# Actions to be executed when the watchers don't timely report to the sbd",
"# main process or one of the watchers detects that the main process has",
"# died.",
"#",
"# Set timeout-action to comma-separated combination of",
"# noflush|flush plus reboot|crashdump|off.",
"# If just one of both is given the other stays at the default.",
"#",
"# This doesn't affect actions like off, crashdump, reboot explicitly",
"# triggered via message slots.",
"# And it does as well not configure the action a watchdog would",
"# trigger should it run off (there is no generic interface).",
"#",
"SBD_TIMEOUT_ACTION=\"flush,reboot\"",
"",
"## Type: yesno / auto",
"## Default: auto",
"#",
"# If CPUAccounting is enabled default is not to assign any RT-budget",
"# to the system.slice which prevents sbd from running RR-scheduled.",
"#",
"# One way to escape that issue is to move sbd-processes from the",
"# slice they were originally started to root-slice.",
"# Of course starting sbd in a certain slice might be intentional.",
"# Thus in auto-mode sbd will check if the slice has RT-budget assigned.",
"# If that is the case sbd will stay in that slice while it will",
"# be moved to root-slice otherwise.",
"#",
"SBD_MOVE_TO_ROOT_CGROUP=auto",
"",
"## Type: yesno",
"## Default: yes",
"#",
"# If resource startup syncing is enabled then pacemakerd is",
"# gonna wait to be pinged via IPC before it starts resources.",
"# On shutdown pacemakerd is going to wait in a state where it",
"# has cleanly shutdown resources till sbd fetches that state.",
"#",
"# The default is set when building SBD and Pacemaker from source.",
"# Going for 'no' is safer if it can't be assured that SBD and",
"# Pacemaker installed do both support the synchronization feature.",
"# When going with 'yes' - also using package dependencies to",
"# assure SBD & Pacemaker both support the synchronization",
"# feature and are assuming the same default - an SBD configuration",
"# inherited via an upgrade doesn't have to be altered to still",
"# benefit from the new feature.",
"#",
"SBD_SYNC_RESOURCE_STARTUP=yes",
"",
"## Type: string",
"## Default: \"\"",
"#",
"# Additional options for starting sbd",
"#",
"SBD_OPTS=\"-n localhost\""
]
},
"changed": false
}
TASK [Print SBD config lines] **************************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:49
Monday 29 July 2024 08:19:08 -0400 (0:00:00.022) 0:01:14.356 ***********
ok: [managed_node1] => {
"__test_sbd_config_lines": [
"#",
"# Ansible managed",
"#",
"# system_role:ha_cluster",
"",
"",
"",
"## Type: string",
"## Default: \"\"",
"#",
"# SBD_DEVICE specifies the devices to use for exchanging sbd messages",
"# and to monitor. If specifying more than one path, use \";\" as",
"# separator.",
"#",
"SBD_DEVICE=\"\"",
"",
"## Type: yesno",
"## Default: yes",
"#",
"# Whether to enable the pacemaker integration.",
"#",
"SBD_PACEMAKER=yes",
"",
"## Type: always / clean",
"## Default: always",
"#",
"# Specify the start mode for sbd. Setting this to \"clean\" will only",
"# allow sbd to start if it was not previously fenced. See the -S option",
"# in the man page.",
"#",
"SBD_STARTMODE=\"always\"",
"",
"## Type: yesno / integer",
"## Default: no",
"#",
"# Whether to delay after starting sbd on boot for \"msgwait\" seconds.",
"# This may be necessary if your cluster nodes reboot so fast that the",
"# other nodes are still waiting in the fence acknowledgement phase.",
"# This is an occasional issue with virtual machines.",
"#",
"# This can also be enabled by being set to a specific delay value, in",
"# seconds. Sometimes a longer delay than the default, \"msgwait\", is",
"# needed, for example in the cases where it's considered to be safer to",
"# wait longer than:",
"# corosync token timeout + consensus timeout + pcmk_delay_max + msgwait",
"#",
"# Be aware that the special value \"1\" means \"yes\" rather than \"1s\".",
"#",
"# Consider that you might have to adapt the startup-timeout accordingly",
"# if the default isn't sufficient. (TimeoutStartSec for systemd)",
"#",
"# This option may be ignored at a later point, once pacemaker handles",
"# this case better.",
"#",
"SBD_DELAY_START=\"no\"",
"",
"## Type: string",
"## Default: /dev/watchdog",
"#",
"# Watchdog device to use. If set to /dev/null, no watchdog device will",
"# be used.",
"#",
"SBD_WATCHDOG_DEV=\"/dev/watchdog\"",
"",
"## Type: integer",
"## Default: 5",
"#",
"# How long, in seconds, the watchdog will wait before panicking the",
"# node if no-one tickles it.",
"#",
"# This depends mostly on your storage latency; the majority of devices",
"# must be successfully read within this time, or else the node will",
"# self-fence.",
"#",
"# If your sbd device(s) reside on a multipath setup or iSCSI, this",
"# should be the time required to detect a path failure.",
"#",
"# Be aware that watchdog timeout set in the on-disk metadata takes",
"# precedence.",
"#",
"SBD_WATCHDOG_TIMEOUT=\"5\"",
"",
"## Type: string",
"## Default: \"flush,reboot\"",
"#",
"# Actions to be executed when the watchers don't timely report to the sbd",
"# main process or one of the watchers detects that the main process has",
"# died.",
"#",
"# Set timeout-action to comma-separated combination of",
"# noflush|flush plus reboot|crashdump|off.",
"# If just one of both is given the other stays at the default.",
"#",
"# This doesn't affect actions like off, crashdump, reboot explicitly",
"# triggered via message slots.",
"# And it does as well not configure the action a watchdog would",
"# trigger should it run off (there is no generic interface).",
"#",
"SBD_TIMEOUT_ACTION=\"flush,reboot\"",
"",
"## Type: yesno / auto",
"## Default: auto",
"#",
"# If CPUAccounting is enabled default is not to assign any RT-budget",
"# to the system.slice which prevents sbd from running RR-scheduled.",
"#",
"# One way to escape that issue is to move sbd-processes from the",
"# slice they were originally started to root-slice.",
"# Of course starting sbd in a certain slice might be intentional.",
"# Thus in auto-mode sbd will check if the slice has RT-budget assigned.",
"# If that is the case sbd will stay in that slice while it will",
"# be moved to root-slice otherwise.",
"#",
"SBD_MOVE_TO_ROOT_CGROUP=auto",
"",
"## Type: yesno",
"## Default: yes",
"#",
"# If resource startup syncing is enabled then pacemakerd is",
"# gonna wait to be pinged via IPC before it starts resources.",
"# On shutdown pacemakerd is going to wait in a state where it",
"# has cleanly shutdown resources till sbd fetches that state.",
"#",
"# The default is set when building SBD and Pacemaker from source.",
"# Going for 'no' is safer if it can't be assured that SBD and",
"# Pacemaker installed do both support the synchronization feature.",
"# When going with 'yes' - also using package dependencies to",
"# assure SBD & Pacemaker both support the synchronization",
"# feature and are assuming the same default - an SBD configuration",
"# inherited via an upgrade doesn't have to be altered to still",
"# benefit from the new feature.",
"#",
"SBD_SYNC_RESOURCE_STARTUP=yes",
"",
"## Type: string",
"## Default: \"\"",
"#",
"# Additional options for starting sbd",
"#",
"SBD_OPTS=\"-n localhost\""
]
}
TASK [Check SBD config] ********************************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:53
Monday 29 July 2024 08:19:08 -0400 (0:00:00.021) 0:01:14.378 ***********
ok: [managed_node1] => {
"changed": false
}
MSG:
All assertions passed
TASK [Fetch quorum configuration] **********************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:61
Monday 29 July 2024 08:19:08 -0400 (0:00:00.023) 0:01:14.401 ***********
ok: [managed_node1] => {
"changed": false,
"cmd": [
"pcs",
"quorum",
"config"
],
"delta": "0:00:00.502321",
"end": "2024-07-29 08:19:09.611330",
"rc": 0,
"start": "2024-07-29 08:19:09.109009"
}
STDOUT:
Options:
TASK [Check auto tie breaker in quorum configuration] **************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:67
Monday 29 July 2024 08:19:09 -0400 (0:00:00.851) 0:01:15.252 ***********
ok: [managed_node1] => {
"changed": false
}
MSG:
All assertions passed
TASK [Get services status] *****************************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:74
Monday 29 July 2024 08:19:09 -0400 (0:00:00.022) 0:01:15.275 ***********
ok: [managed_node1] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"corosync-notifyd.service": {
"name": "corosync-notifyd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qnetd.service": {
"name": "corosync-qnetd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync.service": {
"name": "corosync.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crm_mon.service": {
"name": "crm_mon.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.fedoraproject.FirewallD1.service": {
"name": "dbus-org.fedoraproject.FirewallD1.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "unknown"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-pvscan@.service": {
"name": "lvm2-pvscan@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"multipathd.service": {
"name": "multipathd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"pacemaker.service": {
"name": "pacemaker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"pcsd-ruby.service": {
"name": "pcsd-ruby.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"pcsd.service": {
"name": "pcsd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sbd.service": {
"name": "sbd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sbd_remote.service": {
"name": "sbd_remote.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "unknown"
}
}
},
"changed": false
}
TASK [Check services status] ***************************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:77
Monday 29 July 2024 08:19:11 -0400 (0:00:01.602) 0:01:16.877 ***********
ok: [managed_node1] => {
"changed": false
}
MSG:
All assertions passed
TASK [Check cluster status] ****************************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:84
Monday 29 July 2024 08:19:11 -0400 (0:00:00.028) 0:01:16.906 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/assert_cluster_running.yml for managed_node1
TASK [Get pcs auth status] *****************************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/assert_cluster_running.yml:3
Monday 29 July 2024 08:19:11 -0400 (0:00:00.031) 0:01:16.937 ***********
ok: [managed_node1] => {
"changed": false,
"cmd": [
"pcs",
"status",
"pcsd"
],
"delta": "0:00:00.529330",
"end": "2024-07-29 08:19:12.173620",
"rc": 0,
"start": "2024-07-29 08:19:11.644290"
}
STDOUT:
localhost: Online
TASK [Check pcs auth] **********************************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/assert_cluster_running.yml:9
Monday 29 July 2024 08:19:12 -0400 (0:00:00.879) 0:01:17.817 ***********
ok: [managed_node1] => {
"changed": false
}
MSG:
All assertions passed
TASK [Get cluster status] ******************************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/assert_cluster_running.yml:14
Monday 29 July 2024 08:19:12 -0400 (0:00:00.020) 0:01:17.837 ***********
ok: [managed_node1] => {
"changed": false,
"cmd": [
"pcs",
"status"
],
"delta": "0:00:00.623292",
"end": "2024-07-29 08:19:13.171912",
"rc": 0,
"start": "2024-07-29 08:19:12.548620"
}
STDOUT:
Cluster name: test-cluster
Cluster Summary:
* Stack: corosync (Pacemaker is running)
* Current DC: localhost (version 2.1.7-5.el8-0f7f88312) - partition with quorum
* Last updated: Mon Jul 29 08:19:13 2024 on localhost
* Last change: Mon Jul 29 08:19:05 2024 by root via root on localhost
* 1 node configured
* 0 resource instances configured
Node List:
* Online: [ localhost ]
Full List of Resources:
* No resources
Daemon Status:
corosync: active/enabled
pacemaker: active/enabled
pcsd: active/enabled
sbd: active/enabled
TASK [Check cluster is running] ************************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/assert_cluster_running.yml:20
Monday 29 July 2024 08:19:13 -0400 (0:00:00.977) 0:01:18.815 ***********
ok: [managed_node1] => {
"changed": false
}
MSG:
All assertions passed
TASK [Check firewall and selinux state] ****************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:87
Monday 29 July 2024 08:19:13 -0400 (0:00:00.024) 0:01:18.839 ***********
included: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/check_firewall_selinux.yml for managed_node1
TASK [Check firewall service status] *******************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/check_firewall_selinux.yml:6
Monday 29 July 2024 08:19:13 -0400 (0:00:00.034) 0:01:18.873 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Check firewall port status] **********************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/check_firewall_selinux.yml:12
Monday 29 July 2024 08:19:13 -0400 (0:00:00.019) 0:01:18.893 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Get associated selinux ports] ********************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/check_firewall_selinux.yml:25
Monday 29 July 2024 08:19:13 -0400 (0:00:00.018) 0:01:18.912 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Check associated selinux ports] ******************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/check_firewall_selinux.yml:33
Monday 29 July 2024 08:19:13 -0400 (0:00:00.024) 0:01:18.936 ***********
skipping: [managed_node1] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Clean up test environment for SBD] ***************************************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:91
Monday 29 July 2024 08:19:13 -0400 (0:00:00.027) 0:01:18.963 ***********
TASK [fedora.linux_system_roles.ha_cluster : Unmount SBD devices] **************
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_cleanup_sbd.yml:9
Monday 29 July 2024 08:19:13 -0400 (0:00:00.092) 0:01:19.056 ***********
changed: [managed_node1] => {
"changed": true,
"cmd": [
"losetup",
"-d",
"/dev/loop0"
],
"delta": "0:00:00.009032",
"end": "2024-07-29 08:19:13.794746",
"rc": 0,
"start": "2024-07-29 08:19:13.785714"
}
TASK [fedora.linux_system_roles.ha_cluster : Delete backing files for SBD devices] ***
task path: /tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_cleanup_sbd.yml:14
Monday 29 July 2024 08:19:13 -0400 (0:00:00.386) 0:01:19.443 ***********
ok: [managed_node1] => {
"changed": false,
"path": "{'changed': True, 'path': '/tmp/ansible.qz3pd4ht_ha_cluster_tests', 'uid': 0, 'gid': 0, 'owner': 'root', 'group': 'root', 'mode': '0600', 'state': 'file', 'secontext': 'unconfined_u:object_r:user_tmp_t:s0', 'size': 0, 'failed': False}",
"state": "absent"
}
META: ran handlers
META: ran handlers
PLAY RECAP *********************************************************************
managed_node1 : ok=117 changed=28 unreachable=0 failed=0 skipped=84 rescued=0 ignored=0
Monday 29 July 2024 08:19:14 -0400 (0:00:00.369) 0:01:19.812 ***********
===============================================================================
fedora.linux_system_roles.ha_cluster : Wait for the cluster to fully start and form membership -- 26.71s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:75
fedora.linux_system_roles.ha_cluster : Install role essential packages --- 2.60s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:11
fedora.linux_system_roles.ha_cluster : Install cluster packages --------- 2.49s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:44
fedora.linux_system_roles.ha_cluster : Stop cluster daemons to reload configuration --- 1.85s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:24
fedora.linux_system_roles.ha_cluster : Populate service facts ----------- 1.69s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml:3
fedora.linux_system_roles.ha_cluster : Get services status - detect corosync-qdevice --- 1.62s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:21
Get services status ----------------------------------------------------- 1.60s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:74
fedora.linux_system_roles.ha_cluster : Get services status - detect SBD --- 1.59s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_common/cluster-enable-disable.yml:16
fedora.linux_system_roles.ha_cluster : Get services status - detect pacemaker --- 1.55s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:135
fedora.linux_system_roles.ha_cluster : Write CIB configuration ---------- 1.39s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:33
fedora.linux_system_roles.ha_cluster : Remove qnetd configuration ------- 1.37s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml:3
fedora.linux_system_roles.ha_cluster : Start pcsd with updated config files and configure it to start on boot --- 1.15s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:88
fedora.linux_system_roles.ha_cluster : Remove CIB tempfiles ------------- 1.05s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:284
Get cluster status ------------------------------------------------------ 0.98s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/assert_cluster_running.yml:14
Gathering Facts --------------------------------------------------------- 0.96s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:9
fedora.linux_system_roles.ha_cluster : Enable or disable configured cluster services on boot --- 0.96s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_common/cluster-enable-disable.yml:3
fedora.linux_system_roles.ha_cluster : Set stonith-watchdog-timeout cluster property in CIB --- 0.94s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:142
fedora.linux_system_roles.ha_cluster : Start corosync ------------------- 0.89s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:47
fedora.linux_system_roles.ha_cluster : Check pcs auth status ------------ 0.89s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-auth.yml:3
Get pcs auth status ----------------------------------------------------- 0.88s
/tmp/tmp.iREXviIN7I/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/assert_cluster_running.yml:3