ansible-playbook 2.9.27
config file = None
configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/local/lib/python3.9/site-packages/ansible
executable location = /usr/local/bin/ansible-playbook
python version = 3.9.19 (main, May 16 2024, 11:40:09) [GCC 8.5.0 20210514 (Red Hat 8.5.0-22)]
No config file found; using defaults
[WARNING]: running playbook inside collection fedora.linux_system_roles
Skipping callback 'actionable', as we already have a stdout callback.
Skipping callback 'counter_enabled', as we already have a stdout callback.
Skipping callback 'debug', as we already have a stdout callback.
Skipping callback 'dense', as we already have a stdout callback.
Skipping callback 'dense', as we already have a stdout callback.
Skipping callback 'full_skip', as we already have a stdout callback.
Skipping callback 'json', as we already have a stdout callback.
Skipping callback 'minimal', as we already have a stdout callback.
Skipping callback 'null', as we already have a stdout callback.
Skipping callback 'oneline', as we already have a stdout callback.
Skipping callback 'selective', as we already have a stdout callback.
Skipping callback 'skippy', as we already have a stdout callback.
Skipping callback 'stderr', as we already have a stdout callback.
Skipping callback 'unixy', as we already have a stdout callback.
Skipping callback 'yaml', as we already have a stdout callback.
PLAYBOOK: tests_sbd_defaults.yml ***********************************************
2 plays in /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml
PLAY [all] *********************************************************************
META: ran handlers
TASK [Include vault variables] *************************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:5
Monday 25 November 2024 18:14:10 -0500 (0:00:00.029) 0:00:00.029 *******
ok: [managed-node2] => {
"ansible_facts": {
"ha_cluster_hacluster_password": {
"__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n31303833633366333561656439323930303361333161363239346166656537323933313436\n3432386236656563343237306335323637396239616230353561330a313731623238393238\n62343064666336643930663239383936616465643134646536656532323461356237646133\n3761616633323839633232353637366266350a313163633236376666653238633435306565\n3264623032333736393535663833\n"
}
},
"ansible_included_var_files": [
"/tmp/ha_cluster-ERh/tests/vars/vault-variables.yml"
],
"changed": false
}
META: ran handlers
META: ran handlers
PLAY [Minimal SBD configuration] ***********************************************
TASK [Gathering Facts] *********************************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:9
Monday 25 November 2024 18:14:10 -0500 (0:00:00.043) 0:00:00.072 *******
ok: [managed-node2]
META: ran handlers
TASK [Set up test environment] *************************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:20
Monday 25 November 2024 18:14:11 -0500 (0:00:01.171) 0:00:01.244 *******
TASK [fedora.linux_system_roles.ha_cluster : Set node name to 'localhost' for single-node clusters] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:9
Monday 25 November 2024 18:14:11 -0500 (0:00:00.074) 0:00:01.318 *******
ok: [managed-node2] => {
"ansible_facts": {
"inventory_hostname": "localhost"
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Ensure facts used by tests] *******
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:14
Monday 25 November 2024 18:14:11 -0500 (0:00:00.054) 0:00:01.373 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Check if system is ostree] ********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:22
Monday 25 November 2024 18:14:11 -0500 (0:00:00.065) 0:00:01.439 *******
ok: [managed-node2] => {
"changed": false,
"stat": {
"exists": false
}
}
TASK [fedora.linux_system_roles.ha_cluster : Set flag to indicate system is ostree] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:27
Monday 25 November 2024 18:14:12 -0500 (0:00:00.536) 0:00:01.976 *******
ok: [managed-node2] => {
"ansible_facts": {
"__ha_cluster_is_ostree": false
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Do not try to enable RHEL repositories] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:32
Monday 25 November 2024 18:14:12 -0500 (0:00:00.050) 0:00:02.027 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Copy nss-altfiles ha_cluster users to /etc/passwd] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:41
Monday 25 November 2024 18:14:12 -0500 (0:00:00.054) 0:00:02.081 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Set up test environment for SBD] *****************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:25
Monday 25 November 2024 18:14:12 -0500 (0:00:00.050) 0:00:02.131 *******
TASK [fedora.linux_system_roles.ha_cluster : Load softdog module for SBD to have at least one watchdog] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup_sbd.yml:9
Monday 25 November 2024 18:14:12 -0500 (0:00:00.044) 0:00:02.176 *******
changed: [managed-node2] => {
"changed": true,
"cmd": [
"modprobe",
"softdog"
],
"delta": "0:00:00.006163",
"end": "2024-11-25 18:14:12.760990",
"rc": 0,
"start": "2024-11-25 18:14:12.754827"
}
TASK [fedora.linux_system_roles.ha_cluster : Create backing files for SBD devices] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup_sbd.yml:15
Monday 25 November 2024 18:14:12 -0500 (0:00:00.582) 0:00:02.759 *******
changed: [managed-node2] => {
"changed": true,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.1j7reysj_ha_cluster_tests",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Initialize backing files for SBD devices] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup_sbd.yml:21
Monday 25 November 2024 18:14:13 -0500 (0:00:00.451) 0:00:03.210 *******
changed: [managed-node2] => {
"changed": true,
"cmd": [
"dd",
"if=/dev/zero",
"of=/tmp/ansible.1j7reysj_ha_cluster_tests",
"bs=1M",
"count=10"
],
"delta": "0:00:00.010830",
"end": "2024-11-25 18:14:13.611141",
"rc": 0,
"start": "2024-11-25 18:14:13.600311"
}
STDERR:
10+0 records in
10+0 records out
10485760 bytes (10 MB, 10 MiB) copied, 0.0045715 s, 2.3 GB/s
TASK [fedora.linux_system_roles.ha_cluster : Mount SBD devices] ****************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup_sbd.yml:26
Monday 25 November 2024 18:14:13 -0500 (0:00:00.397) 0:00:03.607 *******
changed: [managed-node2] => {
"changed": true,
"cmd": [
"losetup",
"--show",
"--find",
"/tmp/ansible.1j7reysj_ha_cluster_tests"
],
"delta": "0:00:00.016865",
"end": "2024-11-25 18:14:14.012302",
"rc": 0,
"start": "2024-11-25 18:14:13.995437"
}
STDOUT:
/dev/loop0
TASK [Ensure SBD config file is not present] ***********************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:30
Monday 25 November 2024 18:14:14 -0500 (0:00:00.406) 0:00:04.014 *******
changed: [managed-node2] => {
"changed": true,
"path": "/etc/sysconfig/sbd",
"state": "absent"
}
TASK [Run HA Cluster role] *****************************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:35
Monday 25 November 2024 18:14:14 -0500 (0:00:00.519) 0:00:04.533 *******
TASK [fedora.linux_system_roles.ha_cluster : Set platform/version specific variables] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:3
Monday 25 November 2024 18:14:14 -0500 (0:00:00.068) 0:00:04.602 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Ensure ansible_facts used by role] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:2
Monday 25 November 2024 18:14:14 -0500 (0:00:00.031) 0:00:04.633 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Check if system is ostree] ********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:10
Monday 25 November 2024 18:14:14 -0500 (0:00:00.049) 0:00:04.682 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Set flag to indicate system is ostree] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:15
Monday 25 November 2024 18:14:14 -0500 (0:00:00.037) 0:00:04.720 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Set platform/version specific variables] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:19
Monday 25 November 2024 18:14:14 -0500 (0:00:00.038) 0:00:04.758 *******
ok: [managed-node2] => (item=RedHat.yml) => {
"ansible_facts": {
"__ha_cluster_cloud_agents_packages": [],
"__ha_cluster_fence_agent_packages_default": "{{ ['fence-agents-all'] + (['fence-virt'] if ansible_architecture == 'x86_64' else []) }}",
"__ha_cluster_fullstack_node_packages": [
"corosync",
"libknet1-plugins-all",
"resource-agents",
"pacemaker"
],
"__ha_cluster_pcs_provider": "pcs-0.10",
"__ha_cluster_qdevice_node_packages": [
"corosync-qdevice",
"bash",
"coreutils",
"curl",
"grep",
"nss-tools",
"openssl",
"sed"
],
"__ha_cluster_repos": [],
"__ha_cluster_role_essential_packages": [
"pcs",
"corosync-qnetd",
"openssl"
],
"__ha_cluster_sbd_packages": [
"sbd"
],
"__ha_cluster_services": [
"corosync",
"corosync-qdevice",
"pacemaker"
]
},
"ansible_included_var_files": [
"/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/RedHat.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "RedHat.yml"
}
skipping: [managed-node2] => (item=CentOS.yml) => {
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS.yml",
"skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_8.yml) => {
"ansible_facts": {
"__ha_cluster_cloud_agents_packages": [
"resource-agents-aliyun",
"resource-agents-gcp",
"fence-agents-aliyun",
"fence-agents-aws",
"fence-agents-azure-arm",
"fence-agents-gce"
],
"__ha_cluster_repos": [
{
"id": "ha",
"name": "HighAvailability"
},
{
"id": "resilientstorage",
"name": "ResilientStorage"
}
]
},
"ansible_included_var_files": [
"/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/CentOS_8.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS_8.yml"
}
ok: [managed-node2] => (item=CentOS_8.yml) => {
"ansible_facts": {
"__ha_cluster_cloud_agents_packages": [
"resource-agents-aliyun",
"resource-agents-gcp",
"fence-agents-aliyun",
"fence-agents-aws",
"fence-agents-azure-arm",
"fence-agents-gce"
],
"__ha_cluster_repos": [
{
"id": "ha",
"name": "HighAvailability"
},
{
"id": "resilientstorage",
"name": "ResilientStorage"
}
]
},
"ansible_included_var_files": [
"/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/CentOS_8.yml"
],
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS_8.yml"
}
TASK [fedora.linux_system_roles.ha_cluster : Set Linux Pacemaker shell specific variables] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:34
Monday 25 November 2024 18:14:14 -0500 (0:00:00.068) 0:00:04.827 *******
ok: [managed-node2] => {
"ansible_facts": {},
"ansible_included_var_files": [
"/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/shell_pcs.yml"
],
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Enable package repositories] ******
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:6
Monday 25 November 2024 18:14:14 -0500 (0:00:00.038) 0:00:04.865 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-package-repositories.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Find platform/version specific tasks to enable repositories] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-package-repositories.yml:3
Monday 25 November 2024 18:14:14 -0500 (0:00:00.022) 0:00:04.887 *******
ok: [managed-node2] => (item=RedHat.yml) => {
"ansible_facts": {
"__ha_cluster_enable_repo_tasks_file": "/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/RedHat.yml"
},
"ansible_loop_var": "item",
"changed": false,
"item": "RedHat.yml"
}
ok: [managed-node2] => (item=CentOS.yml) => {
"ansible_facts": {
"__ha_cluster_enable_repo_tasks_file": "/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml"
},
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS.yml"
}
skipping: [managed-node2] => (item=CentOS_8.yml) => {
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS_8.yml",
"skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS_8.yml) => {
"ansible_loop_var": "item",
"changed": false,
"item": "CentOS_8.yml",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Run platform/version specific tasks to enable repositories] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-package-repositories.yml:21
Monday 25 November 2024 18:14:15 -0500 (0:00:00.064) 0:00:04.952 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : List active CentOS repositories] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml:3
Monday 25 November 2024 18:14:15 -0500 (0:00:00.047) 0:00:04.999 *******
[WARNING]: Consider using the dnf module rather than running 'dnf'. If you
need to use command because dnf is insufficient you can add 'warn: false' to
this command task or set 'command_warnings=False' in ansible.cfg to get rid of
this message.
ok: [managed-node2] => {
"changed": false,
"cmd": [
"dnf",
"repolist"
],
"delta": "0:00:00.242874",
"end": "2024-11-25 18:14:15.606337",
"rc": 0,
"start": "2024-11-25 18:14:15.363463"
}
STDOUT:
repo id repo name
appstream CentOS Stream 8 - AppStream
baseos CentOS Stream 8 - BaseOS
beaker-client Beaker Client - RedHatEnterpriseLinux8
beaker-harness Beaker harness
beakerlib-libraries Copr repo for beakerlib-libraries owned by bgoncalv
copr:copr.devel.redhat.com:lpol:qa-tools Copr repo for qa-tools owned by lpol
epel Extra Packages for Enterprise Linux 8 - x86_64
epel-debuginfo Extra Packages for Enterprise Linux 8 - x86_64 - Debug
epel-source Extra Packages for Enterprise Linux 8 - x86_64 - Source
extras CentOS Stream 8 - Extras
extras-common CentOS Stream 8 - Extras common packages
ha CentOS Stream 8 - HighAvailability
TASK [fedora.linux_system_roles.ha_cluster : Enable CentOS repositories] *******
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml:10
Monday 25 November 2024 18:14:15 -0500 (0:00:00.613) 0:00:05.613 *******
skipping: [managed-node2] => (item={'id': 'ha', 'name': 'HighAvailability'}) => {
"ansible_loop_var": "item",
"changed": false,
"item": {
"id": "ha",
"name": "HighAvailability"
},
"skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item={'id': 'resilientstorage', 'name': 'ResilientStorage'}) => {
"ansible_loop_var": "item",
"changed": false,
"item": {
"id": "resilientstorage",
"name": "ResilientStorage"
},
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Install role essential packages] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:11
Monday 25 November 2024 18:14:15 -0500 (0:00:00.058) 0:00:05.671 *******
ok: [managed-node2] => {
"changed": false,
"rc": 0,
"results": []
}
MSG:
Nothing to do
TASK [fedora.linux_system_roles.ha_cluster : Check and prepare role variables] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:17
Monday 25 November 2024 18:14:18 -0500 (0:00:03.037) 0:00:08.708 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Discover cluster node names] ******
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:3
Monday 25 November 2024 18:14:18 -0500 (0:00:00.058) 0:00:08.767 *******
ok: [managed-node2] => {
"ansible_facts": {
"__ha_cluster_node_name": "localhost"
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Collect cluster node names] *******
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:7
Monday 25 November 2024 18:14:18 -0500 (0:00:00.038) 0:00:08.806 *******
ok: [managed-node2] => {
"ansible_facts": {
"__ha_cluster_all_node_names": [
"localhost"
]
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if ha_cluster_node_options contains unknown or duplicate nodes] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:16
Monday 25 November 2024 18:14:18 -0500 (0:00:00.042) 0:00:08.848 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Extract node options] *************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:30
Monday 25 November 2024 18:14:18 -0500 (0:00:00.037) 0:00:08.885 *******
ok: [managed-node2] => {
"ansible_facts": {
"__ha_cluster_local_node": {}
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if passwords are not specified] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:43
Monday 25 November 2024 18:14:18 -0500 (0:00:00.040) 0:00:08.926 *******
skipping: [managed-node2] => (item=ha_cluster_hacluster_password) => {
"ansible_loop_var": "item",
"changed": false,
"item": "ha_cluster_hacluster_password",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if nodes do not have the same number of SBD devices specified] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:53
Monday 25 November 2024 18:14:19 -0500 (0:00:00.065) 0:00:08.991 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if configuring qnetd on a cluster node] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:69
Monday 25 November 2024 18:14:19 -0500 (0:00:00.043) 0:00:09.035 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if no valid level is specified for a fencing level] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:79
Monday 25 November 2024 18:14:19 -0500 (0:00:00.036) 0:00:09.072 *******
TASK [fedora.linux_system_roles.ha_cluster : Fail if no target is specified for a fencing level] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:87
Monday 25 November 2024 18:14:19 -0500 (0:00:00.039) 0:00:09.111 *******
TASK [fedora.linux_system_roles.ha_cluster : Extract qdevice settings] *********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:101
Monday 25 November 2024 18:14:19 -0500 (0:00:00.046) 0:00:09.158 *******
ok: [managed-node2] => {
"ansible_facts": {
"__ha_cluster_qdevice_host": "",
"__ha_cluster_qdevice_in_use": false,
"__ha_cluster_qdevice_model": "",
"__ha_cluster_qdevice_pcs_address": ""
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Figure out if ATB needs to be enabled for SBD] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:110
Monday 25 November 2024 18:14:19 -0500 (0:00:00.072) 0:00:09.230 *******
ok: [managed-node2] => {
"ansible_facts": {
"__ha_cluster_sbd_needs_atb": false
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if SBD needs ATB enabled and the user configured ATB to be disabled] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:120
Monday 25 November 2024 18:14:19 -0500 (0:00:00.091) 0:00:09.322 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if ha_cluster_pcsd_public_key_src and ha_cluster_pcsd_private_key_src are set along with ha_cluster_pcsd_certificates] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:127
Monday 25 November 2024 18:14:19 -0500 (0:00:00.048) 0:00:09.370 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch pcs capabilities] ***********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:141
Monday 25 November 2024 18:14:19 -0500 (0:00:00.051) 0:00:09.421 *******
ok: [managed-node2] => {
"changed": false,
"cmd": [
"pcs",
"--version",
"--full"
],
"delta": "0:00:00.528014",
"end": "2024-11-25 18:14:20.321903",
"rc": 0,
"start": "2024-11-25 18:14:19.793889"
}
STDOUT:
0.10.18
booth booth.enable-authfile.set booth.enable-authfile.unset cluster.config.backup-local cluster.config.restore-cluster cluster.config.restore-local cluster.config.uuid cluster.create cluster.create.enable cluster.create.local cluster.create.no-keys-sync cluster.create.separated-name-and-address cluster.create.start cluster.create.start.wait cluster.create.transport.knet cluster.create.transport.udp-udpu cluster.create.transport.udp-udpu.no-rrp cluster.destroy cluster.destroy.all cluster.report cluster.verify corosync.authkey.update corosync.config.get corosync.config.get.struct corosync.config.reload corosync.config.sync-to-local-cluster corosync.config.update corosync.link.add corosync.link.remove corosync.link.remove.list corosync.link.update corosync.qdevice corosync.qdevice.model.net corosync.quorum corosync.quorum.device corosync.quorum.device.heuristics corosync.quorum.device.model.net corosync.quorum.device.model.net.options_tls_and_kaptb corosync.quorum.set-expected-votes-runtime corosync.quorum.status corosync.quorum.unblock corosync.totem.block_unlisted_ips corosync.uidgid node.add node.add.enable node.add.separated-name-and-address node.add.start node.add.start.wait node.attributes node.attributes.set-list-for-node node.confirm-off node.fence node.guest node.kill node.maintenance node.maintenance.all node.maintenance.list node.maintenance.wait node.remote node.remote.onfail-demote node.remove node.remove-from-caches node.remove.list node.standby node.standby.all node.standby.list node.standby.wait node.start-stop-enable-disable node.start-stop-enable-disable.all node.start-stop-enable-disable.list node.start-stop-enable-disable.start-wait node.utilization node.utilization.set-list-for-node pcmk.acl.enable-disable pcmk.acl.group pcmk.acl.role pcmk.acl.role.create-with-permissions pcmk.acl.role.delete-with-users-groups pcmk.acl.user pcmk.alert pcmk.cib.checkpoints pcmk.cib.checkpoints.diff pcmk.cib.edit pcmk.cib.get pcmk.cib.get.scope pcmk.cib.roles.promoted-unpromoted pcmk.cib.set pcmk.constraint.colocation.set pcmk.constraint.colocation.set.options pcmk.constraint.colocation.simple pcmk.constraint.colocation.simple.options pcmk.constraint.hide-expired pcmk.constraint.location.simple pcmk.constraint.location.simple.options pcmk.constraint.location.simple.resource-regexp pcmk.constraint.location.simple.rule pcmk.constraint.location.simple.rule.node-attr-type-number pcmk.constraint.location.simple.rule.options pcmk.constraint.location.simple.rule.rule-add-remove pcmk.constraint.no-autocorrect pcmk.constraint.order.set pcmk.constraint.order.set.options pcmk.constraint.order.simple pcmk.constraint.order.simple.options pcmk.constraint.ticket.set pcmk.constraint.ticket.set.options pcmk.constraint.ticket.simple pcmk.constraint.ticket.simple.constraint-id pcmk.properties.cluster pcmk.properties.cluster.config.output-formats pcmk.properties.cluster.defaults pcmk.properties.cluster.describe pcmk.properties.cluster.describe.output-formats pcmk.properties.operation-defaults pcmk.properties.operation-defaults.multiple pcmk.properties.operation-defaults.rule pcmk.properties.operation-defaults.rule-rsc-op pcmk.properties.operation-defaults.rule.hide-expired pcmk.properties.operation-defaults.rule.node-attr-type-number pcmk.properties.resource-defaults pcmk.properties.resource-defaults.multiple pcmk.properties.resource-defaults.rule pcmk.properties.resource-defaults.rule-rsc-op pcmk.properties.resource-defaults.rule.hide-expired pcmk.properties.resource-defaults.rule.node-attr-type-number pcmk.resource.ban-move-clear pcmk.resource.ban-move-clear.bundles pcmk.resource.ban-move-clear.clear-expired pcmk.resource.ban-move-clear.clone pcmk.resource.bundle pcmk.resource.bundle.container-docker pcmk.resource.bundle.container-docker.promoted-max pcmk.resource.bundle.container-podman pcmk.resource.bundle.container-podman.promoted-max pcmk.resource.bundle.container-rkt pcmk.resource.bundle.container-rkt.promoted-max pcmk.resource.bundle.reset pcmk.resource.bundle.wait pcmk.resource.cleanup pcmk.resource.cleanup.one-resource pcmk.resource.cleanup.strict pcmk.resource.clone pcmk.resource.clone.custom-id pcmk.resource.clone.meta-in-create pcmk.resource.clone.wait pcmk.resource.config.output-formats pcmk.resource.create pcmk.resource.create.clone.custom-id pcmk.resource.create.in-existing-bundle pcmk.resource.create.meta pcmk.resource.create.no-master pcmk.resource.create.operations pcmk.resource.create.operations.onfail-demote pcmk.resource.create.promotable pcmk.resource.create.promotable.custom-id pcmk.resource.create.wait pcmk.resource.debug pcmk.resource.delete pcmk.resource.disable.safe pcmk.resource.disable.safe.brief pcmk.resource.disable.safe.tag pcmk.resource.disable.simulate pcmk.resource.disable.simulate.brief pcmk.resource.disable.simulate.tag pcmk.resource.enable-disable pcmk.resource.enable-disable.list pcmk.resource.enable-disable.tag pcmk.resource.enable-disable.wait pcmk.resource.failcount pcmk.resource.group pcmk.resource.group.add-remove-list pcmk.resource.group.wait pcmk.resource.manage-unmanage pcmk.resource.manage-unmanage.list pcmk.resource.manage-unmanage.tag pcmk.resource.manage-unmanage.with-monitor pcmk.resource.move.autoclean pcmk.resource.promotable pcmk.resource.promotable.custom-id pcmk.resource.promotable.meta-in-create pcmk.resource.promotable.wait pcmk.resource.refresh pcmk.resource.refresh.one-resource pcmk.resource.refresh.strict pcmk.resource.relations pcmk.resource.relocate pcmk.resource.restart pcmk.resource.update pcmk.resource.update-meta pcmk.resource.update-meta.list pcmk.resource.update-meta.wait pcmk.resource.update-operations pcmk.resource.update-operations.onfail-demote pcmk.resource.update.meta pcmk.resource.update.operations pcmk.resource.update.operations.onfail-demote pcmk.resource.update.wait pcmk.resource.utilization pcmk.resource.utilization-set-list-for-resource pcmk.stonith.cleanup pcmk.stonith.cleanup.one-resource pcmk.stonith.cleanup.strict pcmk.stonith.create pcmk.stonith.create.in-group pcmk.stonith.create.meta pcmk.stonith.create.operations pcmk.stonith.create.operations.onfail-demote pcmk.stonith.create.wait pcmk.stonith.delete pcmk.stonith.enable-disable pcmk.stonith.enable-disable.list pcmk.stonith.enable-disable.wait pcmk.stonith.history.cleanup pcmk.stonith.history.show pcmk.stonith.history.update pcmk.stonith.levels pcmk.stonith.levels.add-remove-devices-list pcmk.stonith.levels.clear pcmk.stonith.levels.node-attr pcmk.stonith.levels.node-regexp pcmk.stonith.levels.verify pcmk.stonith.refresh pcmk.stonith.refresh.one-resource pcmk.stonith.refresh.strict pcmk.stonith.update pcmk.stonith.update.scsi-devices pcmk.stonith.update.scsi-devices.add-remove pcmk.stonith.update.scsi-devices.mpath pcmk.tag pcmk.tag.resources pcs.auth.client pcs.auth.client.cluster pcs.auth.client.token pcs.auth.deauth-client pcs.auth.deauth-server pcs.auth.no-bidirectional pcs.auth.separated-name-and-address pcs.auth.server.token pcs.cfg-in-file.cib pcs.daemon-ssl-cert.set pcs.daemon-ssl-cert.sync-to-local-cluster pcs.disaster-recovery.essentials pcs.request-timeout resource-agents.describe resource-agents.list resource-agents.list.detailed resource-agents.ocf.version-1-0 resource-agents.ocf.version-1-1 resource-agents.self-validation sbd sbd.option-timeout-action sbd.shared-block-device status.corosync.membership status.pcmk.resources.hide-inactive status.pcmk.resources.id status.pcmk.resources.node status.pcmk.resources.orphaned status.pcmk.xml stonith-agents.describe stonith-agents.list stonith-agents.list.detailed stonith-agents.ocf.version-1-0 stonith-agents.ocf.version-1-1 stonith-agents.self-validation
TASK [fedora.linux_system_roles.ha_cluster : Parse pcs capabilities] ***********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:148
Monday 25 November 2024 18:14:20 -0500 (0:00:00.898) 0:00:10.319 *******
ok: [managed-node2] => {
"ansible_facts": {
"__ha_cluster_pcs_capabilities": [
"booth",
"booth.enable-authfile.set",
"booth.enable-authfile.unset",
"cluster.config.backup-local",
"cluster.config.restore-cluster",
"cluster.config.restore-local",
"cluster.config.uuid",
"cluster.create",
"cluster.create.enable",
"cluster.create.local",
"cluster.create.no-keys-sync",
"cluster.create.separated-name-and-address",
"cluster.create.start",
"cluster.create.start.wait",
"cluster.create.transport.knet",
"cluster.create.transport.udp-udpu",
"cluster.create.transport.udp-udpu.no-rrp",
"cluster.destroy",
"cluster.destroy.all",
"cluster.report",
"cluster.verify",
"corosync.authkey.update",
"corosync.config.get",
"corosync.config.get.struct",
"corosync.config.reload",
"corosync.config.sync-to-local-cluster",
"corosync.config.update",
"corosync.link.add",
"corosync.link.remove",
"corosync.link.remove.list",
"corosync.link.update",
"corosync.qdevice",
"corosync.qdevice.model.net",
"corosync.quorum",
"corosync.quorum.device",
"corosync.quorum.device.heuristics",
"corosync.quorum.device.model.net",
"corosync.quorum.device.model.net.options_tls_and_kaptb",
"corosync.quorum.set-expected-votes-runtime",
"corosync.quorum.status",
"corosync.quorum.unblock",
"corosync.totem.block_unlisted_ips",
"corosync.uidgid",
"node.add",
"node.add.enable",
"node.add.separated-name-and-address",
"node.add.start",
"node.add.start.wait",
"node.attributes",
"node.attributes.set-list-for-node",
"node.confirm-off",
"node.fence",
"node.guest",
"node.kill",
"node.maintenance",
"node.maintenance.all",
"node.maintenance.list",
"node.maintenance.wait",
"node.remote",
"node.remote.onfail-demote",
"node.remove",
"node.remove-from-caches",
"node.remove.list",
"node.standby",
"node.standby.all",
"node.standby.list",
"node.standby.wait",
"node.start-stop-enable-disable",
"node.start-stop-enable-disable.all",
"node.start-stop-enable-disable.list",
"node.start-stop-enable-disable.start-wait",
"node.utilization",
"node.utilization.set-list-for-node",
"pcmk.acl.enable-disable",
"pcmk.acl.group",
"pcmk.acl.role",
"pcmk.acl.role.create-with-permissions",
"pcmk.acl.role.delete-with-users-groups",
"pcmk.acl.user",
"pcmk.alert",
"pcmk.cib.checkpoints",
"pcmk.cib.checkpoints.diff",
"pcmk.cib.edit",
"pcmk.cib.get",
"pcmk.cib.get.scope",
"pcmk.cib.roles.promoted-unpromoted",
"pcmk.cib.set",
"pcmk.constraint.colocation.set",
"pcmk.constraint.colocation.set.options",
"pcmk.constraint.colocation.simple",
"pcmk.constraint.colocation.simple.options",
"pcmk.constraint.hide-expired",
"pcmk.constraint.location.simple",
"pcmk.constraint.location.simple.options",
"pcmk.constraint.location.simple.resource-regexp",
"pcmk.constraint.location.simple.rule",
"pcmk.constraint.location.simple.rule.node-attr-type-number",
"pcmk.constraint.location.simple.rule.options",
"pcmk.constraint.location.simple.rule.rule-add-remove",
"pcmk.constraint.no-autocorrect",
"pcmk.constraint.order.set",
"pcmk.constraint.order.set.options",
"pcmk.constraint.order.simple",
"pcmk.constraint.order.simple.options",
"pcmk.constraint.ticket.set",
"pcmk.constraint.ticket.set.options",
"pcmk.constraint.ticket.simple",
"pcmk.constraint.ticket.simple.constraint-id",
"pcmk.properties.cluster",
"pcmk.properties.cluster.config.output-formats",
"pcmk.properties.cluster.defaults",
"pcmk.properties.cluster.describe",
"pcmk.properties.cluster.describe.output-formats",
"pcmk.properties.operation-defaults",
"pcmk.properties.operation-defaults.multiple",
"pcmk.properties.operation-defaults.rule",
"pcmk.properties.operation-defaults.rule-rsc-op",
"pcmk.properties.operation-defaults.rule.hide-expired",
"pcmk.properties.operation-defaults.rule.node-attr-type-number",
"pcmk.properties.resource-defaults",
"pcmk.properties.resource-defaults.multiple",
"pcmk.properties.resource-defaults.rule",
"pcmk.properties.resource-defaults.rule-rsc-op",
"pcmk.properties.resource-defaults.rule.hide-expired",
"pcmk.properties.resource-defaults.rule.node-attr-type-number",
"pcmk.resource.ban-move-clear",
"pcmk.resource.ban-move-clear.bundles",
"pcmk.resource.ban-move-clear.clear-expired",
"pcmk.resource.ban-move-clear.clone",
"pcmk.resource.bundle",
"pcmk.resource.bundle.container-docker",
"pcmk.resource.bundle.container-docker.promoted-max",
"pcmk.resource.bundle.container-podman",
"pcmk.resource.bundle.container-podman.promoted-max",
"pcmk.resource.bundle.container-rkt",
"pcmk.resource.bundle.container-rkt.promoted-max",
"pcmk.resource.bundle.reset",
"pcmk.resource.bundle.wait",
"pcmk.resource.cleanup",
"pcmk.resource.cleanup.one-resource",
"pcmk.resource.cleanup.strict",
"pcmk.resource.clone",
"pcmk.resource.clone.custom-id",
"pcmk.resource.clone.meta-in-create",
"pcmk.resource.clone.wait",
"pcmk.resource.config.output-formats",
"pcmk.resource.create",
"pcmk.resource.create.clone.custom-id",
"pcmk.resource.create.in-existing-bundle",
"pcmk.resource.create.meta",
"pcmk.resource.create.no-master",
"pcmk.resource.create.operations",
"pcmk.resource.create.operations.onfail-demote",
"pcmk.resource.create.promotable",
"pcmk.resource.create.promotable.custom-id",
"pcmk.resource.create.wait",
"pcmk.resource.debug",
"pcmk.resource.delete",
"pcmk.resource.disable.safe",
"pcmk.resource.disable.safe.brief",
"pcmk.resource.disable.safe.tag",
"pcmk.resource.disable.simulate",
"pcmk.resource.disable.simulate.brief",
"pcmk.resource.disable.simulate.tag",
"pcmk.resource.enable-disable",
"pcmk.resource.enable-disable.list",
"pcmk.resource.enable-disable.tag",
"pcmk.resource.enable-disable.wait",
"pcmk.resource.failcount",
"pcmk.resource.group",
"pcmk.resource.group.add-remove-list",
"pcmk.resource.group.wait",
"pcmk.resource.manage-unmanage",
"pcmk.resource.manage-unmanage.list",
"pcmk.resource.manage-unmanage.tag",
"pcmk.resource.manage-unmanage.with-monitor",
"pcmk.resource.move.autoclean",
"pcmk.resource.promotable",
"pcmk.resource.promotable.custom-id",
"pcmk.resource.promotable.meta-in-create",
"pcmk.resource.promotable.wait",
"pcmk.resource.refresh",
"pcmk.resource.refresh.one-resource",
"pcmk.resource.refresh.strict",
"pcmk.resource.relations",
"pcmk.resource.relocate",
"pcmk.resource.restart",
"pcmk.resource.update",
"pcmk.resource.update-meta",
"pcmk.resource.update-meta.list",
"pcmk.resource.update-meta.wait",
"pcmk.resource.update-operations",
"pcmk.resource.update-operations.onfail-demote",
"pcmk.resource.update.meta",
"pcmk.resource.update.operations",
"pcmk.resource.update.operations.onfail-demote",
"pcmk.resource.update.wait",
"pcmk.resource.utilization",
"pcmk.resource.utilization-set-list-for-resource",
"pcmk.stonith.cleanup",
"pcmk.stonith.cleanup.one-resource",
"pcmk.stonith.cleanup.strict",
"pcmk.stonith.create",
"pcmk.stonith.create.in-group",
"pcmk.stonith.create.meta",
"pcmk.stonith.create.operations",
"pcmk.stonith.create.operations.onfail-demote",
"pcmk.stonith.create.wait",
"pcmk.stonith.delete",
"pcmk.stonith.enable-disable",
"pcmk.stonith.enable-disable.list",
"pcmk.stonith.enable-disable.wait",
"pcmk.stonith.history.cleanup",
"pcmk.stonith.history.show",
"pcmk.stonith.history.update",
"pcmk.stonith.levels",
"pcmk.stonith.levels.add-remove-devices-list",
"pcmk.stonith.levels.clear",
"pcmk.stonith.levels.node-attr",
"pcmk.stonith.levels.node-regexp",
"pcmk.stonith.levels.verify",
"pcmk.stonith.refresh",
"pcmk.stonith.refresh.one-resource",
"pcmk.stonith.refresh.strict",
"pcmk.stonith.update",
"pcmk.stonith.update.scsi-devices",
"pcmk.stonith.update.scsi-devices.add-remove",
"pcmk.stonith.update.scsi-devices.mpath",
"pcmk.tag",
"pcmk.tag.resources",
"pcs.auth.client",
"pcs.auth.client.cluster",
"pcs.auth.client.token",
"pcs.auth.deauth-client",
"pcs.auth.deauth-server",
"pcs.auth.no-bidirectional",
"pcs.auth.separated-name-and-address",
"pcs.auth.server.token",
"pcs.cfg-in-file.cib",
"pcs.daemon-ssl-cert.set",
"pcs.daemon-ssl-cert.sync-to-local-cluster",
"pcs.disaster-recovery.essentials",
"pcs.request-timeout",
"resource-agents.describe",
"resource-agents.list",
"resource-agents.list.detailed",
"resource-agents.ocf.version-1-0",
"resource-agents.ocf.version-1-1",
"resource-agents.self-validation",
"sbd",
"sbd.option-timeout-action",
"sbd.shared-block-device",
"status.corosync.membership",
"status.pcmk.resources.hide-inactive",
"status.pcmk.resources.id",
"status.pcmk.resources.node",
"status.pcmk.resources.orphaned",
"status.pcmk.xml",
"stonith-agents.describe",
"stonith-agents.list",
"stonith-agents.list.detailed",
"stonith-agents.ocf.version-1-0",
"stonith-agents.ocf.version-1-1",
"stonith-agents.self-validation"
],
"__ha_cluster_pcsd_capabilities_available": false
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch pcsd capabilities] **********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:155
Monday 25 November 2024 18:14:20 -0500 (0:00:00.047) 0:00:10.367 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Parse pcsd capabilities] **********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:163
Monday 25 November 2024 18:14:20 -0500 (0:00:00.017) 0:00:10.385 *******
ok: [managed-node2] => {
"ansible_facts": {
"__ha_cluster_pcsd_capabilities": []
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fail if pcs is to old to configure resources and operations defaults] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:172
Monday 25 November 2024 18:14:20 -0500 (0:00:00.045) 0:00:10.431 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Generate a password hash] *********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:26
Monday 25 November 2024 18:14:20 -0500 (0:00:00.045) 0:00:10.476 *******
ok: [managed-node2] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Set hacluster password] ***********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:38
Monday 25 November 2024 18:14:21 -0500 (0:00:00.519) 0:00:10.996 *******
ok: [managed-node2] => {
"append": false,
"changed": false,
"comment": "cluster user",
"group": 189,
"home": "/home/hacluster",
"move_home": false,
"name": "hacluster",
"password": "NOT_LOGGING_PASSWORD",
"shell": "/sbin/nologin",
"state": "present",
"uid": 189
}
TASK [fedora.linux_system_roles.ha_cluster : Configure shell] ******************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:43
Monday 25 November 2024 18:14:21 -0500 (0:00:00.792) 0:00:11.788 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Stop pcsd] ************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:6
Monday 25 November 2024 18:14:21 -0500 (0:00:00.061) 0:00:11.850 *******
changed: [managed-node2] => {
"changed": true,
"name": "pcsd",
"state": "stopped",
"status": {
"ActiveEnterTimestamp": "Mon 2024-11-25 18:11:58 EST",
"ActiveEnterTimestampMonotonic": "496399161",
"ActiveExitTimestamp": "Mon 2024-11-25 18:11:56 EST",
"ActiveExitTimestampMonotonic": "494022669",
"ActiveState": "active",
"After": "systemd-journald.socket sysinit.target basic.target network-online.target system.slice pcsd-ruby.service",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Mon 2024-11-25 18:11:58 EST",
"AssertTimestampMonotonic": "496159249",
"Before": "multi-user.target shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Mon 2024-11-25 18:11:58 EST",
"ConditionTimestampMonotonic": "496159247",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ConsistsOf": "pcsd-ruby.service",
"ControlGroup": "/system.slice/pcsd.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "PCS GUI and remote configuration interface",
"DevicePolicy": "auto",
"Documentation": "man:pcsd(8) man:pcs(8)",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/pcsd (ignore_errors=no)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "28215",
"ExecMainStartTimestamp": "Mon 2024-11-25 18:11:58 EST",
"ExecMainStartTimestampMonotonic": "496160417",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/pcsd ; argv[]=/usr/sbin/pcsd ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/pcsd.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "pcsd.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestamp": "Mon 2024-11-25 18:11:56 EST",
"InactiveEnterTimestampMonotonic": "494055445",
"InactiveExitTimestamp": "Mon 2024-11-25 18:11:58 EST",
"InactiveExitTimestampMonotonic": "496160643",
"InvocationID": "8029376f5ffe4e48916a8016b63580df",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "28215",
"MemoryAccounting": "yes",
"MemoryCurrent": "23638016",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "pcsd.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "sysinit.target network-online.target system.slice pcsd-ruby.service",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Mon 2024-11-25 18:11:58 EST",
"StateChangeTimestampMonotonic": "496399161",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "1",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogTimestamp": "Mon 2024-11-25 18:11:58 EST",
"WatchdogTimestampMonotonic": "496399157",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Regenerate pcsd TLS certificate and key] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:11
Monday 25 November 2024 18:14:22 -0500 (0:00:00.862) 0:00:12.713 *******
skipping: [managed-node2] => (item=/var/lib/pcsd/pcsd.key) => {
"ansible_loop_var": "item",
"changed": false,
"item": "/var/lib/pcsd/pcsd.key",
"skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=/var/lib/pcsd/pcsd.crt) => {
"ansible_loop_var": "item",
"changed": false,
"item": "/var/lib/pcsd/pcsd.crt",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Get the stat of /var/lib/pcsd] ****
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:25
Monday 25 November 2024 18:14:22 -0500 (0:00:00.074) 0:00:12.787 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Allow certmonger to write into pcsd's certificate directory] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:30
Monday 25 November 2024 18:14:22 -0500 (0:00:00.073) 0:00:12.860 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Ensure the name of ha_cluster_pcsd_certificates is /var/lib/pcsd/pcsd; Create certificates using the certificate role] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:37
Monday 25 November 2024 18:14:23 -0500 (0:00:00.082) 0:00:12.942 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Set pcsd's certificate directory back to cluster_var_lib_t] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:49
Monday 25 November 2024 18:14:23 -0500 (0:00:00.053) 0:00:12.996 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute pcsd TLS private key] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:64
Monday 25 November 2024 18:14:23 -0500 (0:00:00.049) 0:00:13.045 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute pcsd TLS certificate] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:71
Monday 25 November 2024 18:14:23 -0500 (0:00:00.045) 0:00:13.091 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute pcs_settings.conf] *****
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:79
Monday 25 November 2024 18:14:23 -0500 (0:00:00.067) 0:00:13.159 *******
ok: [managed-node2] => {
"changed": false,
"checksum": "9bbea634a798cf0976b80cd3c4e34aca6a6a7d44",
"dest": "/var/lib/pcsd/pcs_settings.conf",
"gid": 0,
"group": "root",
"mode": "0644",
"owner": "root",
"path": "/var/lib/pcsd/pcs_settings.conf",
"secontext": "system_u:object_r:cluster_var_lib_t:s0",
"size": 361,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Start pcsd with updated config files and configure it to start on boot] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:88
Monday 25 November 2024 18:14:23 -0500 (0:00:00.747) 0:00:13.906 *******
changed: [managed-node2] => {
"changed": true,
"enabled": true,
"name": "pcsd",
"state": "started",
"status": {
"ActiveEnterTimestamp": "Mon 2024-11-25 18:11:58 EST",
"ActiveEnterTimestampMonotonic": "496399161",
"ActiveExitTimestamp": "Mon 2024-11-25 18:14:22 EST",
"ActiveExitTimestampMonotonic": "640671635",
"ActiveState": "inactive",
"After": "systemd-journald.socket sysinit.target basic.target network-online.target system.slice pcsd-ruby.service",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Mon 2024-11-25 18:11:58 EST",
"AssertTimestampMonotonic": "496159249",
"Before": "multi-user.target shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Mon 2024-11-25 18:11:58 EST",
"ConditionTimestampMonotonic": "496159247",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ConsistsOf": "pcsd-ruby.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "PCS GUI and remote configuration interface",
"DevicePolicy": "auto",
"Documentation": "man:pcsd(8) man:pcs(8)",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/pcsd (ignore_errors=no)",
"ExecMainCode": "1",
"ExecMainExitTimestamp": "Mon 2024-11-25 18:14:22 EST",
"ExecMainExitTimestampMonotonic": "640703775",
"ExecMainPID": "28215",
"ExecMainStartTimestamp": "Mon 2024-11-25 18:11:58 EST",
"ExecMainStartTimestampMonotonic": "496160417",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/pcsd ; argv[]=/usr/sbin/pcsd ; ignore_errors=no ; start_time=[Mon 2024-11-25 18:11:58 EST] ; stop_time=[Mon 2024-11-25 18:14:22 EST] ; pid=28215 ; code=exited ; status=0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/pcsd.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "pcsd.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestamp": "Mon 2024-11-25 18:14:22 EST",
"InactiveEnterTimestampMonotonic": "640703890",
"InactiveExitTimestamp": "Mon 2024-11-25 18:11:58 EST",
"InactiveExitTimestampMonotonic": "496160643",
"InvocationID": "8029376f5ffe4e48916a8016b63580df",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "pcsd.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "sysinit.target network-online.target system.slice pcsd-ruby.service",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Mon 2024-11-25 18:14:22 EST",
"StateChangeTimestampMonotonic": "640703890",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Configure firewall] ***************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:49
Monday 25 November 2024 18:14:25 -0500 (0:00:01.286) 0:00:15.193 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/firewall.yml for managed-node2
TASK [Ensure the service and the ports status with the firewall role] **********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/firewall.yml:3
Monday 25 November 2024 18:14:25 -0500 (0:00:00.050) 0:00:15.244 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure selinux] ****************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:52
Monday 25 November 2024 18:14:25 -0500 (0:00:00.056) 0:00:15.300 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Populate service facts] ***********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml:3
Monday 25 November 2024 18:14:25 -0500 (0:00:00.071) 0:00:15.372 *******
ok: [managed-node2] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"corosync-notifyd.service": {
"name": "corosync-notifyd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qdevice.service": {
"name": "corosync-qdevice.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qnetd.service": {
"name": "corosync-qnetd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync.service": {
"name": "corosync.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crm_mon.service": {
"name": "crm_mon.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "unknown"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-pvscan@.service": {
"name": "lvm2-pvscan@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"multipathd.service": {
"name": "multipathd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"pacemaker.service": {
"name": "pacemaker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"pcsd-ruby.service": {
"name": "pcsd-ruby.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"pcsd.service": {
"name": "pcsd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sbd.service": {
"name": "sbd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"sbd_remote.service": {
"name": "sbd_remote.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "unknown"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Set the fence-virt/fence-agents port to _ha_cluster_selinux] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml:12
Monday 25 November 2024 18:14:27 -0500 (0:00:01.948) 0:00:17.320 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Get associated selinux ports] *****
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml:20
Monday 25 November 2024 18:14:27 -0500 (0:00:00.068) 0:00:17.388 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Add the high-availability service ports to _ha_cluster_selinux] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml:28
Monday 25 November 2024 18:14:27 -0500 (0:00:00.083) 0:00:17.472 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Ensure the service and the ports status with the selinux role] ***********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml:35
Monday 25 November 2024 18:14:27 -0500 (0:00:00.091) 0:00:17.563 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Install cluster packages] *********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:58
Monday 25 November 2024 18:14:27 -0500 (0:00:00.071) 0:00:17.635 *******
ok: [managed-node2] => {
"changed": false,
"rc": 0,
"results": []
}
MSG:
Nothing to do
TASK [fedora.linux_system_roles.ha_cluster : Distribute fence-virt authkey] ****
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:64
Monday 25 November 2024 18:14:30 -0500 (0:00:02.982) 0:00:20.617 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/distribute-fence-virt-key.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Create /etc/cluster directory] ****
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/distribute-fence-virt-key.yml:3
Monday 25 November 2024 18:14:30 -0500 (0:00:00.054) 0:00:20.672 *******
ok: [managed-node2] => {
"changed": false,
"gid": 0,
"group": "root",
"mode": "0755",
"owner": "root",
"path": "/etc/cluster",
"secontext": "unconfined_u:object_r:etc_t:s0",
"size": 27,
"state": "directory",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Get fence_xvm.key] ****************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/distribute-fence-virt-key.yml:9
Monday 25 November 2024 18:14:31 -0500 (0:00:00.394) 0:00:21.066 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Check if key exists on the controller fence_xvm.key] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:16
Monday 25 November 2024 18:14:31 -0500 (0:00:00.077) 0:00:21.144 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Slurp existing key from the controller fence_xvm.key] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:21
Monday 25 November 2024 18:14:31 -0500 (0:00:00.031) 0:00:21.176 *******
skipping: [managed-node2] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Use the slurped key from the controller fence_xvm.key] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:29
Monday 25 November 2024 18:14:31 -0500 (0:00:00.031) 0:00:21.207 *******
skipping: [managed-node2] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Generate key using OpenSSL fence_xvm.key] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:44
Monday 25 November 2024 18:14:31 -0500 (0:00:00.037) 0:00:21.244 *******
ok: [managed-node2] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch generated fence_xvm.key] ****
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:50
Monday 25 November 2024 18:14:31 -0500 (0:00:00.445) 0:00:21.690 *******
ok: [managed-node2] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Check if key exists on cluster nodes fence_xvm.key] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:60
Monday 25 November 2024 18:14:31 -0500 (0:00:00.082) 0:00:21.773 *******
ok: [managed-node2] => {
"changed": false,
"stat": {
"atime": 1732576325.8312886,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 8,
"charset": "binary",
"checksum": "12204ed368e235b535288080c3c0ea3075bff545",
"ctime": 1732576325.8392885,
"dev": 51713,
"device_type": 0,
"executable": false,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 278921430,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/octet-stream",
"mode": "0600",
"mtime": 1732576325.5302906,
"nlink": 1,
"path": "/etc/cluster/fence_xvm.key",
"pw_name": "root",
"readable": true,
"rgrp": false,
"roth": false,
"rusr": true,
"size": 512,
"uid": 0,
"version": "2775668573",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": true,
"xgrp": false,
"xoth": false,
"xusr": false
}
}
TASK [fedora.linux_system_roles.ha_cluster : Slurp existing key from cluster nodes fence_xvm.key] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:65
Monday 25 November 2024 18:14:32 -0500 (0:00:00.514) 0:00:22.287 *******
ok: [managed-node2] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Use the slurped key from cluster nodes fence_xvm.key] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:73
Monday 25 November 2024 18:14:32 -0500 (0:00:00.520) 0:00:22.808 *******
ok: [managed-node2 -> managed-node2] => (item=None) => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
ok: [managed-node2 -> {{ item }}] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Send fence_xvm.key to nodes] ******
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/distribute-fence-virt-key.yml:17
Monday 25 November 2024 18:14:32 -0500 (0:00:00.044) 0:00:22.852 *******
ok: [managed-node2] => {
"changed": false,
"checksum": "12204ed368e235b535288080c3c0ea3075bff545",
"dest": "/etc/cluster/fence_xvm.key",
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/etc/cluster/fence_xvm.key",
"secontext": "system_u:object_r:cluster_conf_t:s0",
"size": 512,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Configure SBD] ********************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:69
Monday 25 November 2024 18:14:33 -0500 (0:00:00.662) 0:00:23.515 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Configure watchdog kernel module blocklist] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:10
Monday 25 November 2024 18:14:33 -0500 (0:00:00.057) 0:00:23.572 *******
TASK [fedora.linux_system_roles.ha_cluster : Unload watchdog kernel modules from blocklist] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:22
Monday 25 November 2024 18:14:33 -0500 (0:00:00.024) 0:00:23.596 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure watchdog kernel modules] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:30
Monday 25 November 2024 18:14:33 -0500 (0:00:00.023) 0:00:23.620 *******
TASK [fedora.linux_system_roles.ha_cluster : Load watchdog kernel modules] *****
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:40
Monday 25 November 2024 18:14:33 -0500 (0:00:00.023) 0:00:23.644 *******
TASK [fedora.linux_system_roles.ha_cluster : Probe SBD devices] ****************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:57
Monday 25 November 2024 18:14:33 -0500 (0:00:00.040) 0:00:23.685 *******
TASK [fedora.linux_system_roles.ha_cluster : Initialize SBD devices] ***********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:70
Monday 25 November 2024 18:14:33 -0500 (0:00:00.017) 0:00:23.703 *******
TASK [fedora.linux_system_roles.ha_cluster : Distribute SBD config] ************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:88
Monday 25 November 2024 18:14:33 -0500 (0:00:00.017) 0:00:23.720 *******
changed: [managed-node2] => {
"changed": true,
"checksum": "d5aa0d6d6a774ebaa2eb634f5ef174bb0688c5f8",
"dest": "/etc/sysconfig/sbd",
"gid": 0,
"group": "root",
"md5sum": "6136a912bc0ec0daff3f97cd6fbd3985",
"mode": "0644",
"owner": "root",
"secontext": "system_u:object_r:etc_t:s0",
"size": 4271,
"src": "/root/.ansible/tmp/ansible-tmp-1732576473.820853-22224-259433149589472/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Ensure /etc/systemd/system/sbd.service.d directory exists] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:109
Monday 25 November 2024 18:14:34 -0500 (0:00:00.917) 0:00:24.638 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Override start timeout for SBD] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:117
Monday 25 November 2024 18:14:34 -0500 (0:00:00.025) 0:00:24.663 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Reload systemd service files] *****
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:131
Monday 25 November 2024 18:14:34 -0500 (0:00:00.024) 0:00:24.687 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Get services status - detect pacemaker] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:135
Monday 25 November 2024 18:14:34 -0500 (0:00:00.047) 0:00:24.735 *******
ok: [managed-node2] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"corosync-notifyd.service": {
"name": "corosync-notifyd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qdevice.service": {
"name": "corosync-qdevice.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qnetd.service": {
"name": "corosync-qnetd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync.service": {
"name": "corosync.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crm_mon.service": {
"name": "crm_mon.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "unknown"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-pvscan@.service": {
"name": "lvm2-pvscan@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"multipathd.service": {
"name": "multipathd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"pacemaker.service": {
"name": "pacemaker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"pcsd-ruby.service": {
"name": "pcsd-ruby.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"pcsd.service": {
"name": "pcsd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sbd.service": {
"name": "sbd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"sbd_remote.service": {
"name": "sbd_remote.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "unknown"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Set stonith-watchdog-timeout cluster property in CIB] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:142
Monday 25 November 2024 18:14:36 -0500 (0:00:01.771) 0:00:26.506 *******
changed: [managed-node2] => {
"changed": true,
"cmd": [
"pcs",
"--force",
"--",
"property",
"set",
"stonith-watchdog-timeout="
],
"delta": "0:00:00.624087",
"end": "2024-11-25 18:14:37.579835",
"rc": 0,
"start": "2024-11-25 18:14:36.955748"
}
TASK [fedora.linux_system_roles.ha_cluster : Correct cib.xml ownership] ********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:155
Monday 25 November 2024 18:14:37 -0500 (0:00:01.069) 0:00:27.576 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Clean cib.xml.sig] ****************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:167
Monday 25 November 2024 18:14:37 -0500 (0:00:00.023) 0:00:27.600 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure corosync] ***************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:72
Monday 25 November 2024 18:14:37 -0500 (0:00:00.023) 0:00:27.623 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-corosync.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Create a corosync.conf tempfile] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-corosync.yml:3
Monday 25 November 2024 18:14:37 -0500 (0:00:00.040) 0:00:27.664 *******
changed: [managed-node2] => {
"changed": true,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.z3wn6gyb_ha_cluster_corosync_conf",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Create a corosync.conf file content] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-corosync.yml:16
Monday 25 November 2024 18:14:38 -0500 (0:00:00.452) 0:00:28.117 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cluster-setup-pcs-0.10.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Create a corosync.conf file content using pcs-0.10] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cluster-setup-pcs-0.10.yml:3
Monday 25 November 2024 18:14:38 -0500 (0:00:00.051) 0:00:28.169 *******
changed: [managed-node2] => {
"changed": true,
"cmd": [
"pcs",
"cluster",
"setup",
"--corosync_conf",
"/tmp/ansible.z3wn6gyb_ha_cluster_corosync_conf",
"--overwrite",
"--no-cluster-uuid",
"--",
"test-cluster",
"localhost"
],
"delta": "0:00:00.527598",
"end": "2024-11-25 18:14:39.136611",
"rc": 0,
"start": "2024-11-25 18:14:38.609013"
}
STDOUT:
No addresses specified for host 'localhost', using 'localhost'
TASK [fedora.linux_system_roles.ha_cluster : Add qdevice configuration to corosync.conf] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cluster-setup-pcs-0.10.yml:81
Monday 25 November 2024 18:14:39 -0500 (0:00:00.964) 0:00:29.134 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch created corosync.conf file] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-corosync.yml:19
Monday 25 November 2024 18:14:39 -0500 (0:00:00.018) 0:00:29.152 *******
ok: [managed-node2] => {
"changed": false,
"content": "dG90ZW0gewogICAgdmVyc2lvbjogMgogICAgY2x1c3Rlcl9uYW1lOiB0ZXN0LWNsdXN0ZXIKICAgIHRyYW5zcG9ydDoga25ldAogICAgY3J5cHRvX2NpcGhlcjogYWVzMjU2CiAgICBjcnlwdG9faGFzaDogc2hhMjU2Cn0KCm5vZGVsaXN0IHsKICAgIG5vZGUgewogICAgICAgIHJpbmcwX2FkZHI6IGxvY2FsaG9zdAogICAgICAgIG5hbWU6IGxvY2FsaG9zdAogICAgICAgIG5vZGVpZDogMQogICAgfQp9CgpxdW9ydW0gewogICAgcHJvdmlkZXI6IGNvcm9zeW5jX3ZvdGVxdW9ydW0KfQoKbG9nZ2luZyB7CiAgICB0b19sb2dmaWxlOiB5ZXMKICAgIGxvZ2ZpbGU6IC92YXIvbG9nL2NsdXN0ZXIvY29yb3N5bmMubG9nCiAgICB0b19zeXNsb2c6IHllcwogICAgdGltZXN0YW1wOiBvbgp9Cg==",
"encoding": "base64",
"source": "/tmp/ansible.z3wn6gyb_ha_cluster_corosync_conf"
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute corosync.conf file] ****
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-corosync.yml:26
Monday 25 November 2024 18:14:39 -0500 (0:00:00.380) 0:00:29.533 *******
ok: [managed-node2] => {
"changed": false,
"checksum": "05d2ec2a2bfa233bb2b4ace4aae02b42cafc012b",
"dest": "/etc/corosync/corosync.conf",
"gid": 0,
"group": "root",
"mode": "0644",
"owner": "root",
"path": "/etc/corosync/corosync.conf",
"secontext": "system_u:object_r:etc_t:s0",
"size": 388,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Remove a corosync.conf tempfile] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-corosync.yml:36
Monday 25 November 2024 18:14:40 -0500 (0:00:00.673) 0:00:30.207 *******
changed: [managed-node2] => {
"changed": true,
"path": "/tmp/ansible.z3wn6gyb_ha_cluster_corosync_conf",
"state": "absent"
}
TASK [fedora.linux_system_roles.ha_cluster : Cluster auth] *********************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:75
Monday 25 November 2024 18:14:40 -0500 (0:00:00.408) 0:00:30.615 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-auth.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Check pcs auth status] ************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-auth.yml:3
Monday 25 November 2024 18:14:40 -0500 (0:00:00.046) 0:00:30.662 *******
ok: [managed-node2] => {
"changed": false,
"cmd": [
"pcs",
"status",
"pcsd",
"--",
"localhost"
],
"delta": "0:00:00.547314",
"end": "2024-11-25 18:14:41.603388",
"failed_when_result": false,
"rc": 0,
"start": "2024-11-25 18:14:41.056074"
}
STDOUT:
localhost: Online
TASK [fedora.linux_system_roles.ha_cluster : Run pcs auth] *********************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-auth.yml:23
Monday 25 November 2024 18:14:41 -0500 (0:00:00.940) 0:00:31.602 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute cluster shared keys] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:80
Monday 25 November 2024 18:14:41 -0500 (0:00:00.022) 0:00:31.625 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Get corosync authkey] *************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:3
Monday 25 November 2024 18:14:41 -0500 (0:00:00.073) 0:00:31.698 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Check if key exists on the controller corosync authkey] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:16
Monday 25 November 2024 18:14:41 -0500 (0:00:00.040) 0:00:31.738 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Slurp existing key from the controller corosync authkey] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:21
Monday 25 November 2024 18:14:41 -0500 (0:00:00.025) 0:00:31.764 *******
skipping: [managed-node2] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Use the slurped key from the controller corosync authkey] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:29
Monday 25 November 2024 18:14:41 -0500 (0:00:00.026) 0:00:31.790 *******
skipping: [managed-node2] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Generate key using OpenSSL corosync authkey] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:44
Monday 25 November 2024 18:14:41 -0500 (0:00:00.021) 0:00:31.811 *******
ok: [managed-node2] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch generated corosync authkey] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:50
Monday 25 November 2024 18:14:42 -0500 (0:00:00.372) 0:00:32.184 *******
ok: [managed-node2] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Check if key exists on cluster nodes corosync authkey] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:60
Monday 25 November 2024 18:14:42 -0500 (0:00:00.028) 0:00:32.212 *******
ok: [managed-node2] => {
"changed": false,
"stat": {
"atime": 1732576346.3711505,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 8,
"charset": "binary",
"checksum": "eb49e67cb3348b74d287fd55cc703ac7968ef42a",
"ctime": 1732576335.749222,
"dev": 51713,
"device_type": 0,
"executable": false,
"exists": true,
"gid": 0,
"gr_name": "root",
"inode": 337641660,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/octet-stream",
"mode": "0400",
"mtime": 1732576335.4232242,
"nlink": 1,
"path": "/etc/corosync/authkey",
"pw_name": "root",
"readable": true,
"rgrp": false,
"roth": false,
"rusr": true,
"size": 256,
"uid": 0,
"version": "4250681068",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": false,
"xgrp": false,
"xoth": false,
"xusr": false
}
}
TASK [fedora.linux_system_roles.ha_cluster : Slurp existing key from cluster nodes corosync authkey] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:65
Monday 25 November 2024 18:14:42 -0500 (0:00:00.389) 0:00:32.601 *******
ok: [managed-node2] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Use the slurped key from cluster nodes corosync authkey] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:73
Monday 25 November 2024 18:14:43 -0500 (0:00:00.387) 0:00:32.989 *******
ok: [managed-node2 -> managed-node2] => (item=None) => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
ok: [managed-node2 -> {{ item }}] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute corosync authkey] ******
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:11
Monday 25 November 2024 18:14:43 -0500 (0:00:00.054) 0:00:33.043 *******
ok: [managed-node2] => {
"changed": false,
"checksum": "eb49e67cb3348b74d287fd55cc703ac7968ef42a",
"dest": "/etc/corosync/authkey",
"gid": 0,
"group": "root",
"mode": "0400",
"owner": "root",
"path": "/etc/corosync/authkey",
"secontext": "system_u:object_r:etc_t:s0",
"size": 256,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Get pacemaker authkey] ************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:20
Monday 25 November 2024 18:14:43 -0500 (0:00:00.721) 0:00:33.765 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Check if key exists on the controller pacemaker authkey] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:16
Monday 25 November 2024 18:14:43 -0500 (0:00:00.054) 0:00:33.819 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Slurp existing key from the controller pacemaker authkey] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:21
Monday 25 November 2024 18:14:43 -0500 (0:00:00.026) 0:00:33.846 *******
skipping: [managed-node2] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Use the slurped key from the controller pacemaker authkey] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:29
Monday 25 November 2024 18:14:43 -0500 (0:00:00.028) 0:00:33.875 *******
skipping: [managed-node2] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Generate key using OpenSSL pacemaker authkey] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:44
Monday 25 November 2024 18:14:43 -0500 (0:00:00.023) 0:00:33.898 *******
ok: [managed-node2] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch generated pacemaker authkey] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:50
Monday 25 November 2024 18:14:44 -0500 (0:00:00.367) 0:00:34.266 *******
ok: [managed-node2] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Check if key exists on cluster nodes pacemaker authkey] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:60
Monday 25 November 2024 18:14:44 -0500 (0:00:00.037) 0:00:34.304 *******
ok: [managed-node2] => {
"changed": false,
"stat": {
"atime": 1732576337.5122101,
"attr_flags": "",
"attributes": [],
"block_size": 4096,
"blocks": 8,
"charset": "binary",
"checksum": "1cf9ffcfa079756e0819fd6a61663cee769a2173",
"ctime": 1732576337.52121,
"dev": 51713,
"device_type": 0,
"executable": false,
"exists": true,
"gid": 189,
"gr_name": "haclient",
"inode": 352321689,
"isblk": false,
"ischr": false,
"isdir": false,
"isfifo": false,
"isgid": false,
"islnk": false,
"isreg": true,
"issock": false,
"isuid": false,
"mimetype": "application/octet-stream",
"mode": "0400",
"mtime": 1732576337.243212,
"nlink": 1,
"path": "/etc/pacemaker/authkey",
"pw_name": "hacluster",
"readable": true,
"rgrp": false,
"roth": false,
"rusr": true,
"size": 256,
"uid": 189,
"version": "3256605700",
"wgrp": false,
"woth": false,
"writeable": true,
"wusr": false,
"xgrp": false,
"xoth": false,
"xusr": false
}
}
TASK [fedora.linux_system_roles.ha_cluster : Slurp existing key from cluster nodes pacemaker authkey] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:65
Monday 25 November 2024 18:14:44 -0500 (0:00:00.392) 0:00:34.696 *******
ok: [managed-node2] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Use the slurped key from cluster nodes pacemaker authkey] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/presharedkey.yml:73
Monday 25 November 2024 18:14:45 -0500 (0:00:00.476) 0:00:35.172 *******
ok: [managed-node2 -> managed-node2] => (item=None) => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
ok: [managed-node2 -> {{ item }}] => {
"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Distribute pacemaker authkey] *****
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:28
Monday 25 November 2024 18:14:45 -0500 (0:00:00.044) 0:00:35.217 *******
ok: [managed-node2] => {
"changed": false,
"checksum": "1cf9ffcfa079756e0819fd6a61663cee769a2173",
"dest": "/etc/pacemaker/authkey",
"gid": 189,
"group": "haclient",
"mode": "0400",
"owner": "hacluster",
"path": "/etc/pacemaker/authkey",
"secontext": "system_u:object_r:etc_t:s0",
"size": 256,
"state": "file",
"uid": 189
}
TASK [fedora.linux_system_roles.ha_cluster : Remove qdevice certificates [CLI]] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:37
Monday 25 November 2024 18:14:45 -0500 (0:00:00.669) 0:00:35.887 *******
changed: [managed-node2] => {
"changed": true,
"cmd": [
"pcs",
"--",
"qdevice",
"net-client",
"destroy"
],
"delta": "0:00:00.524343",
"end": "2024-11-25 18:14:46.830069",
"rc": 0,
"start": "2024-11-25 18:14:46.305726"
}
TASK [fedora.linux_system_roles.ha_cluster : Remove qdevice certificates [API]] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:46
Monday 25 November 2024 18:14:46 -0500 (0:00:00.941) 0:00:36.828 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Obtain and distribute qdevice certificates [CLI]] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:54
Monday 25 November 2024 18:14:46 -0500 (0:00:00.035) 0:00:36.864 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Obtain and distribute qdevice certificates [API]] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:71
Monday 25 November 2024 18:14:46 -0500 (0:00:00.030) 0:00:36.894 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Enable or disable cluster services on boot] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:86
Monday 25 November 2024 18:14:46 -0500 (0:00:00.027) 0:00:36.922 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_common/cluster-enable-disable.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Enable or disable configured cluster services on boot] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_common/cluster-enable-disable.yml:3
Monday 25 November 2024 18:14:47 -0500 (0:00:00.084) 0:00:37.006 *******
ok: [managed-node2] => (item=corosync) => {
"ansible_loop_var": "item",
"changed": false,
"enabled": true,
"item": "corosync",
"name": "corosync",
"status": {
"ActiveEnterTimestamp": "Mon 2024-11-25 18:12:26 EST",
"ActiveEnterTimestampMonotonic": "524759296",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "basic.target network-online.target system.slice sysinit.target systemd-journald.socket",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Mon 2024-11-25 18:12:26 EST",
"AssertTimestampMonotonic": "524368790",
"Before": "shutdown.target pacemaker.service multi-user.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Mon 2024-11-25 18:12:26 EST",
"ConditionTimestampMonotonic": "524368764",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroup": "/system.slice/corosync.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Corosync Cluster Engine",
"DevicePolicy": "auto",
"Documentation": "man:corosync man:corosync.conf man:corosync_overview",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/corosync (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "33019",
"ExecMainStartTimestamp": "Mon 2024-11-25 18:12:26 EST",
"ExecMainStartTimestampMonotonic": "524369819",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/corosync ; argv[]=/usr/sbin/corosync -f $COROSYNC_OPTIONS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStop": "{ path=/usr/sbin/corosync-cfgtool ; argv[]=/usr/sbin/corosync-cfgtool -H --force ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/corosync.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "corosync.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Mon 2024-11-25 18:12:26 EST",
"InactiveExitTimestampMonotonic": "524369864",
"InvocationID": "b24f1700db54448c8eb016620b3b0872",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "33019",
"MemoryAccounting": "yes",
"MemoryCurrent": "143663104",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "corosync.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"RequiredBy": "pacemaker.service",
"Requires": "network-online.target system.slice sysinit.target",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Mon 2024-11-25 18:12:26 EST",
"StateChangeTimestampMonotonic": "524759296",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "9",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogTimestamp": "Mon 2024-11-25 18:12:26 EST",
"WatchdogTimestampMonotonic": "524759277",
"WatchdogUSec": "0"
}
}
skipping: [managed-node2] => (item=corosync-qdevice) => {
"ansible_loop_var": "item",
"changed": false,
"item": "corosync-qdevice",
"skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=pacemaker) => {
"ansible_loop_var": "item",
"changed": false,
"enabled": true,
"item": "pacemaker",
"name": "pacemaker",
"status": {
"ActiveEnterTimestamp": "Mon 2024-11-25 18:12:27 EST",
"ActiveEnterTimestampMonotonic": "525906359",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "corosync.service network.target resource-agents-deps.target system.slice basic.target rsyslog.service dbus.service time-sync.target sysinit.target systemd-journald.socket",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Mon 2024-11-25 18:12:27 EST",
"AssertTimestampMonotonic": "525904900",
"Before": "shutdown.target multi-user.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Mon 2024-11-25 18:12:27 EST",
"ConditionTimestampMonotonic": "525904898",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroup": "/system.slice/pacemaker.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Pacemaker High Availability Cluster Manager",
"DevicePolicy": "auto",
"Documentation": "man:pacemakerd https://clusterlabs.org/pacemaker/doc/",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/sbd (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "33282",
"ExecMainStartTimestamp": "Mon 2024-11-25 18:12:27 EST",
"ExecMainStartTimestampMonotonic": "525906316",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/pacemakerd ; argv[]=/usr/sbin/pacemakerd ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/pacemaker.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "pacemaker.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Mon 2024-11-25 18:12:27 EST",
"InactiveExitTimestampMonotonic": "525906359",
"InvocationID": "2eec08d1914a4d5fa07ef89596dcdfcd",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "process",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "33282",
"MemoryAccounting": "yes",
"MemoryCurrent": "56164352",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "pacemaker.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "corosync.service system.slice sysinit.target",
"Restart": "on-failure",
"RestartUSec": "1s",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "no",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "25s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Mon 2024-11-25 18:12:27 EST",
"StateChangeTimestampMonotonic": "525906359",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "7",
"TasksMax": "infinity",
"TimeoutStartUSec": "1min",
"TimeoutStopUSec": "30min",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "simple",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"Wants": "dbus.service resource-agents-deps.target",
"WatchdogTimestamp": "Mon 2024-11-25 18:12:27 EST",
"WatchdogTimestampMonotonic": "525906357",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Get services status - detect SBD] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_common/cluster-enable-disable.yml:16
Monday 25 November 2024 18:14:48 -0500 (0:00:01.084) 0:00:38.090 *******
ok: [managed-node2] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"corosync-notifyd.service": {
"name": "corosync-notifyd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qdevice.service": {
"name": "corosync-qdevice.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qnetd.service": {
"name": "corosync-qnetd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync.service": {
"name": "corosync.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crm_mon.service": {
"name": "crm_mon.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "unknown"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-pvscan@.service": {
"name": "lvm2-pvscan@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"multipathd.service": {
"name": "multipathd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"pacemaker.service": {
"name": "pacemaker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"pcsd-ruby.service": {
"name": "pcsd-ruby.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"pcsd.service": {
"name": "pcsd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sbd.service": {
"name": "sbd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"sbd_remote.service": {
"name": "sbd_remote.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "unknown"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Enable or disable SBD] ************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_common/cluster-enable-disable.yml:19
Monday 25 November 2024 18:14:49 -0500 (0:00:01.756) 0:00:39.847 *******
changed: [managed-node2] => {
"changed": true,
"enabled": true,
"name": "sbd",
"status": {
"ActiveEnterTimestampMonotonic": "0",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "inactive",
"After": "systemd-journald.socket sysinit.target iscsi.service system.slice basic.target systemd-modules-load.service",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "no",
"AssertTimestampMonotonic": "0",
"Before": "shutdown.target dlm.service pacemaker.service",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "no",
"CanStop": "no",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "no",
"ConditionTimestampMonotonic": "0",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Shared-storage based fencing daemon",
"DevicePolicy": "auto",
"Documentation": "man:sbd(8)",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/sbd (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "0",
"ExecMainStartTimestampMonotonic": "0",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/sbd ; argv[]=/usr/sbin/sbd $SBD_OPTS -p /run/sbd.pid watch ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStop": "{ path=/usr/bin/kill ; argv[]=/usr/bin/kill -TERM $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/sbd.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "sbd.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestampMonotonic": "0",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "sbd.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "none",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PIDFile": "/run/sbd.pid",
"PartOf": "corosync.service",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "yes",
"RefuseManualStop": "yes",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "system.slice sysinit.target",
"Restart": "on-abort",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "inherit",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestampMonotonic": "0",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "forking",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "disabled",
"UtmpMode": "init",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Start the cluster and reload corosync.conf] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:89
Monday 25 November 2024 18:14:50 -0500 (0:00:00.675) 0:00:40.522 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Get services status - detect corosync-qdevice] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:21
Monday 25 November 2024 18:14:50 -0500 (0:00:00.057) 0:00:40.580 *******
ok: [managed-node2] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"corosync-notifyd.service": {
"name": "corosync-notifyd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qdevice.service": {
"name": "corosync-qdevice.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qnetd.service": {
"name": "corosync-qnetd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync.service": {
"name": "corosync.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crm_mon.service": {
"name": "crm_mon.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "unknown"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-pvscan@.service": {
"name": "lvm2-pvscan@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"multipathd.service": {
"name": "multipathd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"pacemaker.service": {
"name": "pacemaker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"pcsd-ruby.service": {
"name": "pcsd-ruby.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"pcsd.service": {
"name": "pcsd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sbd.service": {
"name": "sbd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"sbd_remote.service": {
"name": "sbd_remote.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "unknown"
}
}
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Stop cluster daemons to reload configuration] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:24
Monday 25 November 2024 18:14:52 -0500 (0:00:01.769) 0:00:42.349 *******
changed: [managed-node2] => (item=pacemaker) => {
"ansible_loop_var": "item",
"changed": true,
"item": "pacemaker",
"name": "pacemaker",
"state": "stopped",
"status": {
"ActiveEnterTimestamp": "Mon 2024-11-25 18:12:27 EST",
"ActiveEnterTimestampMonotonic": "525906359",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "basic.target system.slice time-sync.target corosync.service sysinit.target rsyslog.service dbus.service systemd-journald.socket sbd.service network.target resource-agents-deps.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Mon 2024-11-25 18:12:27 EST",
"AssertTimestampMonotonic": "525904900",
"Before": "multi-user.target shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Mon 2024-11-25 18:12:27 EST",
"ConditionTimestampMonotonic": "525904898",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlGroup": "/system.slice/pacemaker.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Pacemaker High Availability Cluster Manager",
"DevicePolicy": "auto",
"Documentation": "man:pacemakerd https://clusterlabs.org/pacemaker/doc/",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/sbd (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "33282",
"ExecMainStartTimestamp": "Mon 2024-11-25 18:12:27 EST",
"ExecMainStartTimestampMonotonic": "525906316",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/pacemakerd ; argv[]=/usr/sbin/pacemakerd ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/pacemaker.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "pacemaker.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Mon 2024-11-25 18:12:27 EST",
"InactiveExitTimestampMonotonic": "525906359",
"InvocationID": "2eec08d1914a4d5fa07ef89596dcdfcd",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "process",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "33282",
"MemoryAccounting": "yes",
"MemoryCurrent": "56160256",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "pacemaker.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "sbd.service corosync.service sysinit.target system.slice",
"Restart": "on-failure",
"RestartUSec": "1s",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "no",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "25s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Mon 2024-11-25 18:12:27 EST",
"StateChangeTimestampMonotonic": "525906359",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "7",
"TasksMax": "infinity",
"TimeoutStartUSec": "1min",
"TimeoutStopUSec": "30min",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "simple",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"Wants": "dbus.service resource-agents-deps.target",
"WatchdogTimestamp": "Mon 2024-11-25 18:12:27 EST",
"WatchdogTimestampMonotonic": "525906357",
"WatchdogUSec": "0"
}
}
ok: [managed-node2] => (item=corosync-qdevice) => {
"ansible_loop_var": "item",
"changed": false,
"item": "corosync-qdevice",
"name": "corosync-qdevice",
"state": "stopped",
"status": {
"ActiveEnterTimestampMonotonic": "0",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "inactive",
"After": "system.slice -.mount corosync.service systemd-journald.socket sysinit.target basic.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "no",
"AssertTimestampMonotonic": "0",
"Before": "shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "no",
"ConditionTimestampMonotonic": "0",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Corosync Qdevice daemon",
"DevicePolicy": "auto",
"Documentation": "man:corosync-qdevice",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/corosync-qdevice (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "0",
"ExecMainStartTimestampMonotonic": "0",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/corosync-qdevice ; argv[]=/usr/sbin/corosync-qdevice -f $COROSYNC_QDEVICE_OPTIONS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/corosync-qdevice.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "corosync-qdevice.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestampMonotonic": "0",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "corosync-qdevice.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "-.mount corosync.service sysinit.target system.slice",
"RequiresMountsFor": "/run/corosync-qdevice",
"Restart": "on-failure",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectory": "corosync-qdevice",
"RuntimeDirectoryMode": "0770",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestampMonotonic": "0",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "disabled",
"UtmpMode": "init",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
changed: [managed-node2] => (item=corosync) => {
"ansible_loop_var": "item",
"changed": true,
"item": "corosync",
"name": "corosync",
"state": "stopped",
"status": {
"ActiveEnterTimestamp": "Mon 2024-11-25 18:12:26 EST",
"ActiveEnterTimestampMonotonic": "524759296",
"ActiveExitTimestampMonotonic": "0",
"ActiveState": "active",
"After": "systemd-journald.socket sysinit.target system.slice network-online.target basic.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Mon 2024-11-25 18:12:26 EST",
"AssertTimestampMonotonic": "524368790",
"Before": "pacemaker.service shutdown.target multi-user.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Mon 2024-11-25 18:12:26 EST",
"ConditionTimestampMonotonic": "524368764",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ConsistsOf": "sbd.service",
"ControlGroup": "/system.slice/corosync.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Corosync Cluster Engine",
"DevicePolicy": "auto",
"Documentation": "man:corosync man:corosync.conf man:corosync_overview",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/corosync (ignore_errors=yes)",
"ExecMainCode": "0",
"ExecMainExitTimestampMonotonic": "0",
"ExecMainPID": "33019",
"ExecMainStartTimestamp": "Mon 2024-11-25 18:12:26 EST",
"ExecMainStartTimestampMonotonic": "524369819",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/corosync ; argv[]=/usr/sbin/corosync -f $COROSYNC_OPTIONS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"ExecStop": "{ path=/usr/sbin/corosync-cfgtool ; argv[]=/usr/sbin/corosync-cfgtool -H --force ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/corosync.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "corosync.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestampMonotonic": "0",
"InactiveExitTimestamp": "Mon 2024-11-25 18:12:26 EST",
"InactiveExitTimestampMonotonic": "524369864",
"InvocationID": "b24f1700db54448c8eb016620b3b0872",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "33019",
"MemoryAccounting": "yes",
"MemoryCurrent": "119164928",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "corosync.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"RequiredBy": "pacemaker.service",
"Requires": "network-online.target system.slice sysinit.target sbd.service",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Mon 2024-11-25 18:12:26 EST",
"StateChangeTimestampMonotonic": "524759296",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "running",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "9",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogTimestamp": "Mon 2024-11-25 18:12:26 EST",
"WatchdogTimestampMonotonic": "524759277",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Start corosync] *******************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:47
Monday 25 November 2024 18:14:54 -0500 (0:00:02.440) 0:00:44.790 *******
changed: [managed-node2] => {
"changed": true,
"name": "corosync",
"state": "started",
"status": {
"ActiveEnterTimestamp": "Mon 2024-11-25 18:12:26 EST",
"ActiveEnterTimestampMonotonic": "524759296",
"ActiveExitTimestamp": "Mon 2024-11-25 18:14:53 EST",
"ActiveExitTimestampMonotonic": "671964122",
"ActiveState": "inactive",
"After": "systemd-journald.socket sysinit.target system.slice network-online.target basic.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Mon 2024-11-25 18:12:26 EST",
"AssertTimestampMonotonic": "524368790",
"Before": "pacemaker.service shutdown.target multi-user.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Mon 2024-11-25 18:12:26 EST",
"ConditionTimestampMonotonic": "524368764",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ConsistsOf": "sbd.service",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Corosync Cluster Engine",
"DevicePolicy": "auto",
"Documentation": "man:corosync man:corosync.conf man:corosync_overview",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/corosync (ignore_errors=yes)",
"ExecMainCode": "1",
"ExecMainExitTimestamp": "Mon 2024-11-25 18:14:54 EST",
"ExecMainExitTimestampMonotonic": "672783672",
"ExecMainPID": "33019",
"ExecMainStartTimestamp": "Mon 2024-11-25 18:12:26 EST",
"ExecMainStartTimestampMonotonic": "524369819",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/corosync ; argv[]=/usr/sbin/corosync -f $COROSYNC_OPTIONS ; ignore_errors=no ; start_time=[Mon 2024-11-25 18:12:26 EST] ; stop_time=[Mon 2024-11-25 18:14:54 EST] ; pid=33019 ; code=exited ; status=0 }",
"ExecStop": "{ path=/usr/sbin/corosync-cfgtool ; argv[]=/usr/sbin/corosync-cfgtool -H --force ; ignore_errors=no ; start_time=[Mon 2024-11-25 18:14:53 EST] ; stop_time=[Mon 2024-11-25 18:14:53 EST] ; pid=50250 ; code=exited ; status=0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/corosync.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "corosync.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestamp": "Mon 2024-11-25 18:14:54 EST",
"InactiveEnterTimestampMonotonic": "672783795",
"InactiveExitTimestamp": "Mon 2024-11-25 18:12:26 EST",
"InactiveExitTimestampMonotonic": "524369864",
"InvocationID": "b24f1700db54448c8eb016620b3b0872",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "control-group",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "corosync.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"RequiredBy": "pacemaker.service",
"Requires": "network-online.target system.slice sysinit.target sbd.service",
"Restart": "no",
"RestartUSec": "100ms",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "yes",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "10s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Mon 2024-11-25 18:14:54 EST",
"StateChangeTimestampMonotonic": "672783795",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "22405",
"TimeoutStartUSec": "1min 30s",
"TimeoutStopUSec": "1min 30s",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "notify",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Reload corosync configuration] ****
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:56
Monday 25 November 2024 18:14:55 -0500 (0:00:00.929) 0:00:45.720 *******
ok: [managed-node2] => {
"changed": false,
"cmd": [
"corosync-cfgtool",
"-R"
],
"delta": "0:00:00.012699",
"end": "2024-11-25 18:14:56.107553",
"rc": 0,
"start": "2024-11-25 18:14:56.094854"
}
STDOUT:
Reloading corosync.conf...
Done
TASK [fedora.linux_system_roles.ha_cluster : Start corosync-qdevice] ***********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:63
Monday 25 November 2024 18:14:56 -0500 (0:00:00.389) 0:00:46.109 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Start pacemaker] ******************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:69
Monday 25 November 2024 18:14:56 -0500 (0:00:00.021) 0:00:46.130 *******
changed: [managed-node2] => {
"changed": true,
"name": "pacemaker",
"state": "started",
"status": {
"ActiveEnterTimestamp": "Mon 2024-11-25 18:12:27 EST",
"ActiveEnterTimestampMonotonic": "525906359",
"ActiveExitTimestamp": "Mon 2024-11-25 18:14:52 EST",
"ActiveExitTimestampMonotonic": "670889548",
"ActiveState": "inactive",
"After": "basic.target system.slice time-sync.target corosync.service sysinit.target rsyslog.service dbus.service systemd-journald.socket sbd.service network.target resource-agents-deps.target",
"AllowIsolate": "no",
"AllowedCPUs": "",
"AllowedMemoryNodes": "",
"AmbientCapabilities": "",
"AssertResult": "yes",
"AssertTimestamp": "Mon 2024-11-25 18:12:27 EST",
"AssertTimestampMonotonic": "525904900",
"Before": "multi-user.target shutdown.target",
"BlockIOAccounting": "no",
"BlockIOWeight": "[not set]",
"CPUAccounting": "no",
"CPUAffinity": "",
"CPUAffinityFromNUMA": "no",
"CPUQuotaPerSecUSec": "infinity",
"CPUQuotaPeriodUSec": "infinity",
"CPUSchedulingPolicy": "0",
"CPUSchedulingPriority": "0",
"CPUSchedulingResetOnFork": "no",
"CPUShares": "[not set]",
"CPUUsageNSec": "[not set]",
"CPUWeight": "[not set]",
"CacheDirectoryMode": "0755",
"CanFreeze": "yes",
"CanIsolate": "no",
"CanReload": "no",
"CanStart": "yes",
"CanStop": "yes",
"CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
"CollectMode": "inactive",
"ConditionResult": "yes",
"ConditionTimestamp": "Mon 2024-11-25 18:12:27 EST",
"ConditionTimestampMonotonic": "525904898",
"ConfigurationDirectoryMode": "0755",
"Conflicts": "shutdown.target",
"ControlPID": "0",
"DefaultDependencies": "yes",
"DefaultMemoryLow": "0",
"DefaultMemoryMin": "0",
"Delegate": "no",
"Description": "Pacemaker High Availability Cluster Manager",
"DevicePolicy": "auto",
"Documentation": "man:pacemakerd https://clusterlabs.org/pacemaker/doc/",
"DynamicUser": "no",
"EffectiveCPUs": "",
"EffectiveMemoryNodes": "",
"EnvironmentFiles": "/etc/sysconfig/sbd (ignore_errors=yes)",
"ExecMainCode": "1",
"ExecMainExitTimestamp": "Mon 2024-11-25 18:14:52 EST",
"ExecMainExitTimestampMonotonic": "670938227",
"ExecMainPID": "33282",
"ExecMainStartTimestamp": "Mon 2024-11-25 18:12:27 EST",
"ExecMainStartTimestampMonotonic": "525906316",
"ExecMainStatus": "0",
"ExecStart": "{ path=/usr/sbin/pacemakerd ; argv[]=/usr/sbin/pacemakerd ; ignore_errors=no ; start_time=[Mon 2024-11-25 18:12:27 EST] ; stop_time=[Mon 2024-11-25 18:14:52 EST] ; pid=33282 ; code=exited ; status=0 }",
"FailureAction": "none",
"FileDescriptorStoreMax": "0",
"FragmentPath": "/usr/lib/systemd/system/pacemaker.service",
"FreezerState": "running",
"GID": "[not set]",
"GuessMainPID": "yes",
"IOAccounting": "no",
"IOSchedulingClass": "0",
"IOSchedulingPriority": "0",
"IOWeight": "[not set]",
"IPAccounting": "no",
"IPEgressBytes": "18446744073709551615",
"IPEgressPackets": "18446744073709551615",
"IPIngressBytes": "18446744073709551615",
"IPIngressPackets": "18446744073709551615",
"Id": "pacemaker.service",
"IgnoreOnIsolate": "no",
"IgnoreSIGPIPE": "yes",
"InactiveEnterTimestamp": "Mon 2024-11-25 18:14:52 EST",
"InactiveEnterTimestampMonotonic": "670938292",
"InactiveExitTimestamp": "Mon 2024-11-25 18:12:27 EST",
"InactiveExitTimestampMonotonic": "525906359",
"InvocationID": "2eec08d1914a4d5fa07ef89596dcdfcd",
"JobRunningTimeoutUSec": "infinity",
"JobTimeoutAction": "none",
"JobTimeoutUSec": "infinity",
"KeyringMode": "private",
"KillMode": "process",
"KillSignal": "15",
"LimitAS": "infinity",
"LimitASSoft": "infinity",
"LimitCORE": "infinity",
"LimitCORESoft": "0",
"LimitCPU": "infinity",
"LimitCPUSoft": "infinity",
"LimitDATA": "infinity",
"LimitDATASoft": "infinity",
"LimitFSIZE": "infinity",
"LimitFSIZESoft": "infinity",
"LimitLOCKS": "infinity",
"LimitLOCKSSoft": "infinity",
"LimitMEMLOCK": "65536",
"LimitMEMLOCKSoft": "65536",
"LimitMSGQUEUE": "819200",
"LimitMSGQUEUESoft": "819200",
"LimitNICE": "0",
"LimitNICESoft": "0",
"LimitNOFILE": "262144",
"LimitNOFILESoft": "1024",
"LimitNPROC": "14003",
"LimitNPROCSoft": "14003",
"LimitRSS": "infinity",
"LimitRSSSoft": "infinity",
"LimitRTPRIO": "0",
"LimitRTPRIOSoft": "0",
"LimitRTTIME": "infinity",
"LimitRTTIMESoft": "infinity",
"LimitSIGPENDING": "14003",
"LimitSIGPENDINGSoft": "14003",
"LimitSTACK": "infinity",
"LimitSTACKSoft": "8388608",
"LoadState": "loaded",
"LockPersonality": "no",
"LogLevelMax": "-1",
"LogRateLimitBurst": "0",
"LogRateLimitIntervalUSec": "0",
"LogsDirectoryMode": "0755",
"MainPID": "0",
"MemoryAccounting": "yes",
"MemoryCurrent": "[not set]",
"MemoryDenyWriteExecute": "no",
"MemoryHigh": "infinity",
"MemoryLimit": "infinity",
"MemoryLow": "0",
"MemoryMax": "infinity",
"MemoryMin": "0",
"MemorySwapMax": "infinity",
"MountAPIVFS": "no",
"MountFlags": "",
"NFileDescriptorStore": "0",
"NRestarts": "0",
"NUMAMask": "",
"NUMAPolicy": "n/a",
"Names": "pacemaker.service",
"NeedDaemonReload": "no",
"Nice": "0",
"NoNewPrivileges": "no",
"NonBlocking": "no",
"NotifyAccess": "main",
"OOMScoreAdjust": "0",
"OnFailureJobMode": "replace",
"PermissionsStartOnly": "no",
"Perpetual": "no",
"PrivateDevices": "no",
"PrivateMounts": "no",
"PrivateNetwork": "no",
"PrivateTmp": "no",
"PrivateUsers": "no",
"ProtectControlGroups": "no",
"ProtectHome": "no",
"ProtectKernelModules": "no",
"ProtectKernelTunables": "no",
"ProtectSystem": "no",
"RefuseManualStart": "no",
"RefuseManualStop": "no",
"RemainAfterExit": "no",
"RemoveIPC": "no",
"Requires": "sbd.service corosync.service sysinit.target system.slice",
"Restart": "on-failure",
"RestartUSec": "1s",
"RestrictNamespaces": "no",
"RestrictRealtime": "no",
"RestrictSUIDSGID": "no",
"Result": "success",
"RootDirectoryStartOnly": "no",
"RuntimeDirectoryMode": "0755",
"RuntimeDirectoryPreserve": "no",
"RuntimeMaxUSec": "infinity",
"SameProcessGroup": "no",
"SecureBits": "0",
"SendSIGHUP": "no",
"SendSIGKILL": "no",
"Slice": "system.slice",
"StandardError": "null",
"StandardInput": "null",
"StandardInputData": "",
"StandardOutput": "journal",
"StartLimitAction": "none",
"StartLimitBurst": "5",
"StartLimitIntervalUSec": "25s",
"StartupBlockIOWeight": "[not set]",
"StartupCPUShares": "[not set]",
"StartupCPUWeight": "[not set]",
"StartupIOWeight": "[not set]",
"StateChangeTimestamp": "Mon 2024-11-25 18:14:52 EST",
"StateChangeTimestampMonotonic": "670938292",
"StateDirectoryMode": "0755",
"StatusErrno": "0",
"StopWhenUnneeded": "no",
"SubState": "dead",
"SuccessAction": "none",
"SyslogFacility": "3",
"SyslogLevel": "6",
"SyslogLevelPrefix": "yes",
"SyslogPriority": "30",
"SystemCallErrorNumber": "0",
"TTYReset": "no",
"TTYVHangup": "no",
"TTYVTDisallocate": "no",
"TasksAccounting": "yes",
"TasksCurrent": "[not set]",
"TasksMax": "infinity",
"TimeoutStartUSec": "1min",
"TimeoutStopUSec": "30min",
"TimerSlackNSec": "50000",
"Transient": "no",
"Type": "simple",
"UID": "[not set]",
"UMask": "0022",
"UnitFilePreset": "disabled",
"UnitFileState": "enabled",
"UtmpMode": "init",
"WantedBy": "multi-user.target",
"Wants": "dbus.service resource-agents-deps.target",
"WatchdogTimestampMonotonic": "0",
"WatchdogUSec": "0"
}
}
TASK [fedora.linux_system_roles.ha_cluster : Wait for the cluster to fully start and form membership] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:75
Monday 25 November 2024 18:14:56 -0500 (0:00:00.558) 0:00:46.689 *******
ok: [managed-node2] => {
"changed": false,
"cmd": [
"pcs",
"cluster",
"start",
"--all",
"--wait"
],
"delta": "0:00:26.397182",
"end": "2024-11-25 18:15:23.482559",
"rc": 0,
"start": "2024-11-25 18:14:57.085377"
}
STDOUT:
localhost: Starting Cluster...
Waiting for node(s) to start...
localhost: Started
TASK [fedora.linux_system_roles.ha_cluster : List pacemaker nodes] *************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:89
Monday 25 November 2024 18:15:23 -0500 (0:00:26.795) 0:01:13.485 *******
ok: [managed-node2] => {
"changed": false,
"cmd": "set -euo pipefail; crm_mon -X | xmllint --xpath '/crm_mon/nodes/node/@name' - | sed -E 's/\\s*name=\"([^\"]+)\"\\s*/\\1\\n/g'\n",
"delta": "0:00:00.021168",
"end": "2024-11-25 18:15:23.869896",
"rc": 0,
"start": "2024-11-25 18:15:23.848728"
}
STDOUT:
localhost
TASK [fedora.linux_system_roles.ha_cluster : Purge removed nodes from pacemaker's cache] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:111
Monday 25 November 2024 18:15:23 -0500 (0:00:00.399) 0:01:13.884 *******
skipping: [managed-node2] => (item=localhost) => {
"ansible_loop_var": "item",
"changed": false,
"item": "localhost",
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Create and push CIB] **************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:92
Monday 25 November 2024 18:15:23 -0500 (0:00:00.034) 0:01:13.918 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Create a tempfile for original CIB] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:10
Monday 25 November 2024 18:15:24 -0500 (0:00:00.093) 0:01:14.011 *******
changed: [managed-node2] => {
"changed": true,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.zxfxgk9g_ha_cluster_original_cib_xml",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Create a tempfile for new CIB] ****
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:18
Monday 25 November 2024 18:15:24 -0500 (0:00:00.358) 0:01:14.370 *******
changed: [managed-node2] => {
"changed": true,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.uui8fwqi_ha_cluster_cib_xml",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Fetch CIB configuration] **********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:26
Monday 25 November 2024 18:15:24 -0500 (0:00:00.361) 0:01:14.732 *******
ok: [managed-node2] => {
"changed": false,
"cmd": [
"cibadmin",
"--query"
],
"delta": "0:00:00.013173",
"end": "2024-11-25 18:15:25.101474",
"rc": 0,
"start": "2024-11-25 18:15:25.088301"
}
STDOUT:
TASK [fedora.linux_system_roles.ha_cluster : Write CIB configuration] **********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:33
Monday 25 November 2024 18:15:25 -0500 (0:00:00.369) 0:01:15.101 *******
changed: [managed-node2] => (item=/tmp/ansible.uui8fwqi_ha_cluster_cib_xml) => {
"ansible_loop_var": "item",
"changed": true,
"checksum": "367e2839a39494fbeb07358c95705faf9ee93017",
"dest": "/tmp/ansible.uui8fwqi_ha_cluster_cib_xml",
"gid": 0,
"group": "root",
"item": "/tmp/ansible.uui8fwqi_ha_cluster_cib_xml",
"md5sum": "b168460298998949e19a7b2d84b6f461",
"mode": "0600",
"owner": "root",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 17372,
"src": "/root/.ansible/tmp/ansible-tmp-1732576525.2130787-23891-119167325844844/source",
"state": "file",
"uid": 0
}
changed: [managed-node2] => (item=/tmp/ansible.zxfxgk9g_ha_cluster_original_cib_xml) => {
"ansible_loop_var": "item",
"changed": true,
"checksum": "367e2839a39494fbeb07358c95705faf9ee93017",
"dest": "/tmp/ansible.zxfxgk9g_ha_cluster_original_cib_xml",
"gid": 0,
"group": "root",
"item": "/tmp/ansible.zxfxgk9g_ha_cluster_original_cib_xml",
"md5sum": "b168460298998949e19a7b2d84b6f461",
"mode": "0600",
"owner": "root",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 17372,
"src": "/root/.ansible/tmp/ansible-tmp-1732576525.887831-23891-206316025985099/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Purge new CIB configuration] ******
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:51
Monday 25 November 2024 18:15:26 -0500 (0:00:01.394) 0:01:16.495 *******
changed: [managed-node2] => {
"changed": true,
"cmd": [
"cibadmin",
"--force",
"--delete-all",
"--xpath",
"/cib/configuration/*[not(\n self::crm_config or\n self::nodes or\n self::resources or\n self::constraints\n)] | /cib/configuration/*[self::resources or self::constraints]/* | /cib/configuration/nodes/*/* | /cib/configuration/crm_config//nvpair[not(\n @name=\"cluster-infrastructure\" or\n @name=\"cluster-name\" or\n @name=\"dc-version\" or\n @name=\"have-watchdog\" or\n @name=\"last-lrm-refresh\" or\n @name=\"stonith-watchdog-timeout\"\n)]"
],
"delta": "0:00:00.024883",
"end": "2024-11-25 18:15:26.888004",
"rc": 0,
"start": "2024-11-25 18:15:26.863121"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure cluster properties] *****
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:85
Monday 25 November 2024 18:15:26 -0500 (0:00:00.396) 0:01:16.891 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure node attributes] ********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:95
Monday 25 November 2024 18:15:26 -0500 (0:00:00.023) 0:01:16.915 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure node utilization] *******
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:102
Monday 25 November 2024 18:15:26 -0500 (0:00:00.022) 0:01:16.937 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure resource defaults] ******
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:115
Monday 25 November 2024 18:15:27 -0500 (0:00:00.020) 0:01:16.958 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure resource operation defaults] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:127
Monday 25 November 2024 18:15:27 -0500 (0:00:00.024) 0:01:16.982 *******
TASK [fedora.linux_system_roles.ha_cluster : Extract primitive to bundle mapping] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:137
Monday 25 November 2024 18:15:27 -0500 (0:00:00.017) 0:01:17.000 *******
ok: [managed-node2] => {
"ansible_facts": {
"__ha_cluster_primitive_bundle_map": {}
},
"changed": false
}
TASK [fedora.linux_system_roles.ha_cluster : Configure cluster bundle resources] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:141
Monday 25 November 2024 18:15:27 -0500 (0:00:00.024) 0:01:17.024 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure cluster resources] ******
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:147
Monday 25 November 2024 18:15:27 -0500 (0:00:00.017) 0:01:17.042 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure cluster resource groups] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:154
Monday 25 November 2024 18:15:27 -0500 (0:00:00.017) 0:01:17.059 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure cluster resource clones] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:160
Monday 25 November 2024 18:15:27 -0500 (0:00:00.019) 0:01:17.078 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure stonith levels] *********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:167
Monday 25 November 2024 18:15:27 -0500 (0:00:00.016) 0:01:17.095 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure resource location constraints] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:175
Monday 25 November 2024 18:15:27 -0500 (0:00:00.017) 0:01:17.113 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure resource colocation constraints] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:182
Monday 25 November 2024 18:15:27 -0500 (0:00:00.016) 0:01:17.130 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure resource set colocation constraints] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:190
Monday 25 November 2024 18:15:27 -0500 (0:00:00.020) 0:01:17.150 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure resource order constraints] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:200
Monday 25 November 2024 18:15:27 -0500 (0:00:00.018) 0:01:17.169 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure resource set order constraints] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:208
Monday 25 November 2024 18:15:27 -0500 (0:00:00.029) 0:01:17.198 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure resource ticket constraints] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:218
Monday 25 November 2024 18:15:27 -0500 (0:00:00.029) 0:01:17.227 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure resource set ticket constraints] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:226
Monday 25 November 2024 18:15:27 -0500 (0:00:00.033) 0:01:17.261 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure acls] *******************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:236
Monday 25 November 2024 18:15:27 -0500 (0:00:00.041) 0:01:17.303 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cib-acls.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Configure ACL roles] **************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cib-acls.yml:3
Monday 25 November 2024 18:15:27 -0500 (0:00:00.048) 0:01:17.351 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure ACL users] **************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cib-acls.yml:28
Monday 25 November 2024 18:15:27 -0500 (0:00:00.022) 0:01:17.373 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure ACL groups] *************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cib-acls.yml:44
Monday 25 November 2024 18:15:27 -0500 (0:00:00.026) 0:01:17.400 *******
TASK [fedora.linux_system_roles.ha_cluster : Configure alerts] *****************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:241
Monday 25 November 2024 18:15:27 -0500 (0:00:00.018) 0:01:17.418 *******
TASK [fedora.linux_system_roles.ha_cluster : Create a tempfile for CIB diff] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:249
Monday 25 November 2024 18:15:27 -0500 (0:00:00.021) 0:01:17.439 *******
changed: [managed-node2] => {
"changed": true,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.4ie8rjdy_ha_cluster_cib_diff",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Compare new and original CIB] *****
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:257
Monday 25 November 2024 18:15:27 -0500 (0:00:00.373) 0:01:17.813 *******
changed: [managed-node2] => {
"changed": true,
"cmd": [
"crm_diff",
"--no-version",
"--original",
"/tmp/ansible.zxfxgk9g_ha_cluster_original_cib_xml",
"--new",
"/tmp/ansible.uui8fwqi_ha_cluster_cib_xml"
],
"delta": "0:00:00.006766",
"end": "2024-11-25 18:15:28.188440",
"failed_when_result": false,
"rc": 1,
"start": "2024-11-25 18:15:28.181674"
}
STDOUT:
MSG:
non-zero return code
TASK [fedora.linux_system_roles.ha_cluster : Write CIB diff to its tempfile] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:270
Monday 25 November 2024 18:15:28 -0500 (0:00:00.398) 0:01:18.211 *******
changed: [managed-node2] => {
"changed": true,
"checksum": "3e1c9ebcf787f7b75eaed83a8c33021580e3eea7",
"dest": "/tmp/ansible.4ie8rjdy_ha_cluster_cib_diff",
"gid": 0,
"group": "root",
"md5sum": "4fea5044c9c2853edeb2a78b166ee6be",
"mode": "0600",
"owner": "root",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 5669,
"src": "/root/.ansible/tmp/ansible-tmp-1732576528.323882-24085-250984086554728/source",
"state": "file",
"uid": 0
}
TASK [fedora.linux_system_roles.ha_cluster : Push CIB diff to the cluster if it has any changes] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:281
Monday 25 November 2024 18:15:29 -0500 (0:00:00.771) 0:01:18.983 *******
changed: [managed-node2] => {
"changed": true,
"cmd": [
"cibadmin",
"--verbose",
"--patch",
"--xml-file",
"/tmp/ansible.4ie8rjdy_ha_cluster_cib_diff"
],
"delta": "0:00:00.022588",
"end": "2024-11-25 18:15:29.390106",
"rc": 0,
"start": "2024-11-25 18:15:29.367518"
}
TASK [fedora.linux_system_roles.ha_cluster : Remove CIB tempfiles] *************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:290
Monday 25 November 2024 18:15:29 -0500 (0:00:00.444) 0:01:19.428 *******
changed: [managed-node2] => (item={'changed': True, 'path': '/tmp/ansible.uui8fwqi_ha_cluster_cib_xml', 'uid': 0, 'gid': 0, 'owner': 'root', 'group': 'root', 'mode': '0600', 'state': 'file', 'secontext': 'unconfined_u:object_r:user_tmp_t:s0', 'size': 0, 'failed': False}) => {
"ansible_loop_var": "item",
"changed": true,
"item": {
"changed": true,
"failed": false,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.uui8fwqi_ha_cluster_cib_xml",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
},
"path": "/tmp/ansible.uui8fwqi_ha_cluster_cib_xml",
"state": "absent"
}
changed: [managed-node2] => (item={'changed': True, 'path': '/tmp/ansible.zxfxgk9g_ha_cluster_original_cib_xml', 'uid': 0, 'gid': 0, 'owner': 'root', 'group': 'root', 'mode': '0600', 'state': 'file', 'secontext': 'unconfined_u:object_r:user_tmp_t:s0', 'size': 0, 'failed': False}) => {
"ansible_loop_var": "item",
"changed": true,
"item": {
"changed": true,
"failed": false,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.zxfxgk9g_ha_cluster_original_cib_xml",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
},
"path": "/tmp/ansible.zxfxgk9g_ha_cluster_original_cib_xml",
"state": "absent"
}
changed: [managed-node2] => (item={'changed': True, 'path': '/tmp/ansible.4ie8rjdy_ha_cluster_cib_diff', 'uid': 0, 'gid': 0, 'owner': 'root', 'group': 'root', 'mode': '0600', 'state': 'file', 'secontext': 'unconfined_u:object_r:user_tmp_t:s0', 'size': 0, 'failed': False}) => {
"ansible_loop_var": "item",
"changed": true,
"item": {
"changed": true,
"failed": false,
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"path": "/tmp/ansible.4ie8rjdy_ha_cluster_cib_diff",
"secontext": "unconfined_u:object_r:user_tmp_t:s0",
"size": 0,
"state": "file",
"uid": 0
},
"path": "/tmp/ansible.4ie8rjdy_ha_cluster_cib_diff",
"state": "absent"
}
TASK [fedora.linux_system_roles.ha_cluster : Remove cluster configuration] *****
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:101
Monday 25 November 2024 18:15:30 -0500 (0:00:01.095) 0:01:20.524 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Remove fence-virt authkey] ********
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:104
Monday 25 November 2024 18:15:30 -0500 (0:00:00.031) 0:01:20.555 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Configure qnetd] ******************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:109
Monday 25 November 2024 18:15:30 -0500 (0:00:00.029) 0:01:20.584 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml for managed-node2
TASK [fedora.linux_system_roles.ha_cluster : Remove qnetd configuration] *******
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml:3
Monday 25 November 2024 18:15:30 -0500 (0:00:00.048) 0:01:20.632 *******
changed: [managed-node2] => {
"changed": true,
"cmd": [
"pcs",
"--force",
"--",
"qdevice",
"destroy",
"net"
],
"delta": "0:00:01.021154",
"end": "2024-11-25 18:15:32.058489",
"rc": 0,
"start": "2024-11-25 18:15:31.037335"
}
STDOUT:
Stopping quorum device...
quorum device stopped
quorum device disabled
Quorum device 'net' configuration files removed
TASK [fedora.linux_system_roles.ha_cluster : Setup qnetd] **********************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml:16
Monday 25 November 2024 18:15:32 -0500 (0:00:01.449) 0:01:22.082 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [fedora.linux_system_roles.ha_cluster : Enable or disable qnetd service on boot] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml:26
Monday 25 November 2024 18:15:32 -0500 (0:00:00.032) 0:01:22.114 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Slurp SBD config file] ***************************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:40
Monday 25 November 2024 18:15:32 -0500 (0:00:00.020) 0:01:22.134 *******
ok: [managed-node2] => {
"changed": false,
"content": "IwojIEFuc2libGUgbWFuYWdlZAojCiMgc3lzdGVtX3JvbGU6aGFfY2x1c3RlcgoKCgojIyBUeXBlOiBzdHJpbmcKIyMgRGVmYXVsdDogIiIKIwojIFNCRF9ERVZJQ0Ugc3BlY2lmaWVzIHRoZSBkZXZpY2VzIHRvIHVzZSBmb3IgZXhjaGFuZ2luZyBzYmQgbWVzc2FnZXMKIyBhbmQgdG8gbW9uaXRvci4gSWYgc3BlY2lmeWluZyBtb3JlIHRoYW4gb25lIHBhdGgsIHVzZSAiOyIgYXMKIyBzZXBhcmF0b3IuCiMKU0JEX0RFVklDRT0iIgoKIyMgVHlwZTogeWVzbm8KIyMgRGVmYXVsdDogeWVzCiMKIyBXaGV0aGVyIHRvIGVuYWJsZSB0aGUgcGFjZW1ha2VyIGludGVncmF0aW9uLgojClNCRF9QQUNFTUFLRVI9eWVzCgojIyBUeXBlOiBhbHdheXMgLyBjbGVhbgojIyBEZWZhdWx0OiBhbHdheXMKIwojIFNwZWNpZnkgdGhlIHN0YXJ0IG1vZGUgZm9yIHNiZC4gU2V0dGluZyB0aGlzIHRvICJjbGVhbiIgd2lsbCBvbmx5CiMgYWxsb3cgc2JkIHRvIHN0YXJ0IGlmIGl0IHdhcyBub3QgcHJldmlvdXNseSBmZW5jZWQuIFNlZSB0aGUgLVMgb3B0aW9uCiMgaW4gdGhlIG1hbiBwYWdlLgojClNCRF9TVEFSVE1PREU9ImFsd2F5cyIKCiMjIFR5cGU6IHllc25vIC8gaW50ZWdlcgojIyBEZWZhdWx0OiBubwojCiMgV2hldGhlciB0byBkZWxheSBhZnRlciBzdGFydGluZyBzYmQgb24gYm9vdCBmb3IgIm1zZ3dhaXQiIHNlY29uZHMuCiMgVGhpcyBtYXkgYmUgbmVjZXNzYXJ5IGlmIHlvdXIgY2x1c3RlciBub2RlcyByZWJvb3Qgc28gZmFzdCB0aGF0IHRoZQojIG90aGVyIG5vZGVzIGFyZSBzdGlsbCB3YWl0aW5nIGluIHRoZSBmZW5jZSBhY2tub3dsZWRnZW1lbnQgcGhhc2UuCiMgVGhpcyBpcyBhbiBvY2Nhc2lvbmFsIGlzc3VlIHdpdGggdmlydHVhbCBtYWNoaW5lcy4KIwojIFRoaXMgY2FuIGFsc28gYmUgZW5hYmxlZCBieSBiZWluZyBzZXQgdG8gYSBzcGVjaWZpYyBkZWxheSB2YWx1ZSwgaW4KIyBzZWNvbmRzLiBTb21ldGltZXMgYSBsb25nZXIgZGVsYXkgdGhhbiB0aGUgZGVmYXVsdCwgIm1zZ3dhaXQiLCBpcwojIG5lZWRlZCwgZm9yIGV4YW1wbGUgaW4gdGhlIGNhc2VzIHdoZXJlIGl0J3MgY29uc2lkZXJlZCB0byBiZSBzYWZlciB0bwojIHdhaXQgbG9uZ2VyIHRoYW46CiMgY29yb3N5bmMgdG9rZW4gdGltZW91dCArIGNvbnNlbnN1cyB0aW1lb3V0ICsgcGNta19kZWxheV9tYXggKyBtc2d3YWl0CiMKIyBCZSBhd2FyZSB0aGF0IHRoZSBzcGVjaWFsIHZhbHVlICIxIiBtZWFucyAieWVzIiByYXRoZXIgdGhhbiAiMXMiLgojCiMgQ29uc2lkZXIgdGhhdCB5b3UgbWlnaHQgaGF2ZSB0byBhZGFwdCB0aGUgc3RhcnR1cC10aW1lb3V0IGFjY29yZGluZ2x5CiMgaWYgdGhlIGRlZmF1bHQgaXNuJ3Qgc3VmZmljaWVudC4gKFRpbWVvdXRTdGFydFNlYyBmb3Igc3lzdGVtZCkKIwojIFRoaXMgb3B0aW9uIG1heSBiZSBpZ25vcmVkIGF0IGEgbGF0ZXIgcG9pbnQsIG9uY2UgcGFjZW1ha2VyIGhhbmRsZXMKIyB0aGlzIGNhc2UgYmV0dGVyLgojClNCRF9ERUxBWV9TVEFSVD0ibm8iCgojIyBUeXBlOiBzdHJpbmcKIyMgRGVmYXVsdDogL2Rldi93YXRjaGRvZwojCiMgV2F0Y2hkb2cgZGV2aWNlIHRvIHVzZS4gSWYgc2V0IHRvIC9kZXYvbnVsbCwgbm8gd2F0Y2hkb2cgZGV2aWNlIHdpbGwKIyBiZSB1c2VkLgojClNCRF9XQVRDSERPR19ERVY9Ii9kZXYvd2F0Y2hkb2ciCgojIyBUeXBlOiBpbnRlZ2VyCiMjIERlZmF1bHQ6IDUKIwojIEhvdyBsb25nLCBpbiBzZWNvbmRzLCB0aGUgd2F0Y2hkb2cgd2lsbCB3YWl0IGJlZm9yZSBwYW5pY2tpbmcgdGhlCiMgbm9kZSBpZiBuby1vbmUgdGlja2xlcyBpdC4KIwojIFRoaXMgZGVwZW5kcyBtb3N0bHkgb24geW91ciBzdG9yYWdlIGxhdGVuY3k7IHRoZSBtYWpvcml0eSBvZiBkZXZpY2VzCiMgbXVzdCBiZSBzdWNjZXNzZnVsbHkgcmVhZCB3aXRoaW4gdGhpcyB0aW1lLCBvciBlbHNlIHRoZSBub2RlIHdpbGwKIyBzZWxmLWZlbmNlLgojCiMgSWYgeW91ciBzYmQgZGV2aWNlKHMpIHJlc2lkZSBvbiBhIG11bHRpcGF0aCBzZXR1cCBvciBpU0NTSSwgdGhpcwojIHNob3VsZCBiZSB0aGUgdGltZSByZXF1aXJlZCB0byBkZXRlY3QgYSBwYXRoIGZhaWx1cmUuCiMKIyBCZSBhd2FyZSB0aGF0IHdhdGNoZG9nIHRpbWVvdXQgc2V0IGluIHRoZSBvbi1kaXNrIG1ldGFkYXRhIHRha2VzCiMgcHJlY2VkZW5jZS4KIwpTQkRfV0FUQ0hET0dfVElNRU9VVD0iNSIKCiMjIFR5cGU6IHN0cmluZwojIyBEZWZhdWx0OiAiZmx1c2gscmVib290IgojCiMgQWN0aW9ucyB0byBiZSBleGVjdXRlZCB3aGVuIHRoZSB3YXRjaGVycyBkb24ndCB0aW1lbHkgcmVwb3J0IHRvIHRoZSBzYmQKIyBtYWluIHByb2Nlc3Mgb3Igb25lIG9mIHRoZSB3YXRjaGVycyBkZXRlY3RzIHRoYXQgdGhlIG1haW4gcHJvY2VzcyBoYXMKIyBkaWVkLgojCiMgU2V0IHRpbWVvdXQtYWN0aW9uIHRvIGNvbW1hLXNlcGFyYXRlZCBjb21iaW5hdGlvbiBvZgojIG5vZmx1c2h8Zmx1c2ggcGx1cyByZWJvb3R8Y3Jhc2hkdW1wfG9mZi4KIyBJZiBqdXN0IG9uZSBvZiBib3RoIGlzIGdpdmVuIHRoZSBvdGhlciBzdGF5cyBhdCB0aGUgZGVmYXVsdC4KIwojIFRoaXMgZG9lc24ndCBhZmZlY3QgYWN0aW9ucyBsaWtlIG9mZiwgY3Jhc2hkdW1wLCByZWJvb3QgZXhwbGljaXRseQojIHRyaWdnZXJlZCB2aWEgbWVzc2FnZSBzbG90cy4KIyBBbmQgaXQgZG9lcyBhcyB3ZWxsIG5vdCBjb25maWd1cmUgdGhlIGFjdGlvbiBhIHdhdGNoZG9nIHdvdWxkCiMgdHJpZ2dlciBzaG91bGQgaXQgcnVuIG9mZiAodGhlcmUgaXMgbm8gZ2VuZXJpYyBpbnRlcmZhY2UpLgojClNCRF9USU1FT1VUX0FDVElPTj0iZmx1c2gscmVib290IgoKIyMgVHlwZTogeWVzbm8gLyBhdXRvCiMjIERlZmF1bHQ6IGF1dG8KIwojIElmIENQVUFjY291bnRpbmcgaXMgZW5hYmxlZCBkZWZhdWx0IGlzIG5vdCB0byBhc3NpZ24gYW55IFJULWJ1ZGdldAojIHRvIHRoZSBzeXN0ZW0uc2xpY2Ugd2hpY2ggcHJldmVudHMgc2JkIGZyb20gcnVubmluZyBSUi1zY2hlZHVsZWQuCiMKIyBPbmUgd2F5IHRvIGVzY2FwZSB0aGF0IGlzc3VlIGlzIHRvIG1vdmUgc2JkLXByb2Nlc3NlcyBmcm9tIHRoZQojIHNsaWNlIHRoZXkgd2VyZSBvcmlnaW5hbGx5IHN0YXJ0ZWQgdG8gcm9vdC1zbGljZS4KIyBPZiBjb3Vyc2Ugc3RhcnRpbmcgc2JkIGluIGEgY2VydGFpbiBzbGljZSBtaWdodCBiZSBpbnRlbnRpb25hbC4KIyBUaHVzIGluIGF1dG8tbW9kZSBzYmQgd2lsbCBjaGVjayBpZiB0aGUgc2xpY2UgaGFzIFJULWJ1ZGdldCBhc3NpZ25lZC4KIyBJZiB0aGF0IGlzIHRoZSBjYXNlIHNiZCB3aWxsIHN0YXkgaW4gdGhhdCBzbGljZSB3aGlsZSBpdCB3aWxsCiMgYmUgbW92ZWQgdG8gcm9vdC1zbGljZSBvdGhlcndpc2UuCiMKU0JEX01PVkVfVE9fUk9PVF9DR1JPVVA9YXV0bwoKIyMgVHlwZTogeWVzbm8KIyMgRGVmYXVsdDogeWVzCiMKIyBJZiByZXNvdXJjZSBzdGFydHVwIHN5bmNpbmcgaXMgZW5hYmxlZCB0aGVuIHBhY2VtYWtlcmQgaXMKIyBnb25uYSB3YWl0IHRvIGJlIHBpbmdlZCB2aWEgSVBDIGJlZm9yZSBpdCBzdGFydHMgcmVzb3VyY2VzLgojIE9uIHNodXRkb3duIHBhY2VtYWtlcmQgaXMgZ29pbmcgdG8gd2FpdCBpbiBhIHN0YXRlIHdoZXJlIGl0CiMgaGFzIGNsZWFubHkgc2h1dGRvd24gcmVzb3VyY2VzIHRpbGwgc2JkIGZldGNoZXMgdGhhdCBzdGF0ZS4KIwojIFRoZSBkZWZhdWx0IGlzIHNldCB3aGVuIGJ1aWxkaW5nIFNCRCBhbmQgUGFjZW1ha2VyIGZyb20gc291cmNlLgojIEdvaW5nIGZvciAnbm8nIGlzIHNhZmVyIGlmIGl0IGNhbid0IGJlIGFzc3VyZWQgdGhhdCBTQkQgYW5kCiMgUGFjZW1ha2VyIGluc3RhbGxlZCBkbyBib3RoIHN1cHBvcnQgdGhlIHN5bmNocm9uaXphdGlvbiBmZWF0dXJlLgojIFdoZW4gZ29pbmcgd2l0aCAneWVzJyAtIGFsc28gdXNpbmcgcGFja2FnZSBkZXBlbmRlbmNpZXMgdG8KIyBhc3N1cmUgU0JEICYgUGFjZW1ha2VyIGJvdGggc3VwcG9ydCB0aGUgc3luY2hyb25pemF0aW9uCiMgZmVhdHVyZSBhbmQgYXJlIGFzc3VtaW5nIHRoZSBzYW1lIGRlZmF1bHQgLSBhbiBTQkQgY29uZmlndXJhdGlvbgojIGluaGVyaXRlZCB2aWEgYW4gdXBncmFkZSBkb2Vzbid0IGhhdmUgdG8gYmUgYWx0ZXJlZCB0byBzdGlsbAojIGJlbmVmaXQgZnJvbSB0aGUgbmV3IGZlYXR1cmUuCiMKU0JEX1NZTkNfUkVTT1VSQ0VfU1RBUlRVUD15ZXMKCiMjIFR5cGU6IHN0cmluZwojIyBEZWZhdWx0OiAiIgojCiMgQWRkaXRpb25hbCBvcHRpb25zIGZvciBzdGFydGluZyBzYmQKIwpTQkRfT1BUUz0iLW4gbG9jYWxob3N0Igo=",
"encoding": "base64",
"source": "/etc/sysconfig/sbd"
}
TASK [Decode SBD config] *******************************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:45
Monday 25 November 2024 18:15:32 -0500 (0:00:00.383) 0:01:22.518 *******
ok: [managed-node2] => {
"ansible_facts": {
"__test_sbd_config_lines": [
"#",
"# Ansible managed",
"#",
"# system_role:ha_cluster",
"",
"",
"",
"## Type: string",
"## Default: \"\"",
"#",
"# SBD_DEVICE specifies the devices to use for exchanging sbd messages",
"# and to monitor. If specifying more than one path, use \";\" as",
"# separator.",
"#",
"SBD_DEVICE=\"\"",
"",
"## Type: yesno",
"## Default: yes",
"#",
"# Whether to enable the pacemaker integration.",
"#",
"SBD_PACEMAKER=yes",
"",
"## Type: always / clean",
"## Default: always",
"#",
"# Specify the start mode for sbd. Setting this to \"clean\" will only",
"# allow sbd to start if it was not previously fenced. See the -S option",
"# in the man page.",
"#",
"SBD_STARTMODE=\"always\"",
"",
"## Type: yesno / integer",
"## Default: no",
"#",
"# Whether to delay after starting sbd on boot for \"msgwait\" seconds.",
"# This may be necessary if your cluster nodes reboot so fast that the",
"# other nodes are still waiting in the fence acknowledgement phase.",
"# This is an occasional issue with virtual machines.",
"#",
"# This can also be enabled by being set to a specific delay value, in",
"# seconds. Sometimes a longer delay than the default, \"msgwait\", is",
"# needed, for example in the cases where it's considered to be safer to",
"# wait longer than:",
"# corosync token timeout + consensus timeout + pcmk_delay_max + msgwait",
"#",
"# Be aware that the special value \"1\" means \"yes\" rather than \"1s\".",
"#",
"# Consider that you might have to adapt the startup-timeout accordingly",
"# if the default isn't sufficient. (TimeoutStartSec for systemd)",
"#",
"# This option may be ignored at a later point, once pacemaker handles",
"# this case better.",
"#",
"SBD_DELAY_START=\"no\"",
"",
"## Type: string",
"## Default: /dev/watchdog",
"#",
"# Watchdog device to use. If set to /dev/null, no watchdog device will",
"# be used.",
"#",
"SBD_WATCHDOG_DEV=\"/dev/watchdog\"",
"",
"## Type: integer",
"## Default: 5",
"#",
"# How long, in seconds, the watchdog will wait before panicking the",
"# node if no-one tickles it.",
"#",
"# This depends mostly on your storage latency; the majority of devices",
"# must be successfully read within this time, or else the node will",
"# self-fence.",
"#",
"# If your sbd device(s) reside on a multipath setup or iSCSI, this",
"# should be the time required to detect a path failure.",
"#",
"# Be aware that watchdog timeout set in the on-disk metadata takes",
"# precedence.",
"#",
"SBD_WATCHDOG_TIMEOUT=\"5\"",
"",
"## Type: string",
"## Default: \"flush,reboot\"",
"#",
"# Actions to be executed when the watchers don't timely report to the sbd",
"# main process or one of the watchers detects that the main process has",
"# died.",
"#",
"# Set timeout-action to comma-separated combination of",
"# noflush|flush plus reboot|crashdump|off.",
"# If just one of both is given the other stays at the default.",
"#",
"# This doesn't affect actions like off, crashdump, reboot explicitly",
"# triggered via message slots.",
"# And it does as well not configure the action a watchdog would",
"# trigger should it run off (there is no generic interface).",
"#",
"SBD_TIMEOUT_ACTION=\"flush,reboot\"",
"",
"## Type: yesno / auto",
"## Default: auto",
"#",
"# If CPUAccounting is enabled default is not to assign any RT-budget",
"# to the system.slice which prevents sbd from running RR-scheduled.",
"#",
"# One way to escape that issue is to move sbd-processes from the",
"# slice they were originally started to root-slice.",
"# Of course starting sbd in a certain slice might be intentional.",
"# Thus in auto-mode sbd will check if the slice has RT-budget assigned.",
"# If that is the case sbd will stay in that slice while it will",
"# be moved to root-slice otherwise.",
"#",
"SBD_MOVE_TO_ROOT_CGROUP=auto",
"",
"## Type: yesno",
"## Default: yes",
"#",
"# If resource startup syncing is enabled then pacemakerd is",
"# gonna wait to be pinged via IPC before it starts resources.",
"# On shutdown pacemakerd is going to wait in a state where it",
"# has cleanly shutdown resources till sbd fetches that state.",
"#",
"# The default is set when building SBD and Pacemaker from source.",
"# Going for 'no' is safer if it can't be assured that SBD and",
"# Pacemaker installed do both support the synchronization feature.",
"# When going with 'yes' - also using package dependencies to",
"# assure SBD & Pacemaker both support the synchronization",
"# feature and are assuming the same default - an SBD configuration",
"# inherited via an upgrade doesn't have to be altered to still",
"# benefit from the new feature.",
"#",
"SBD_SYNC_RESOURCE_STARTUP=yes",
"",
"## Type: string",
"## Default: \"\"",
"#",
"# Additional options for starting sbd",
"#",
"SBD_OPTS=\"-n localhost\""
]
},
"changed": false
}
TASK [Print SBD config lines] **************************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:49
Monday 25 November 2024 18:15:32 -0500 (0:00:00.034) 0:01:22.552 *******
ok: [managed-node2] => {
"__test_sbd_config_lines": [
"#",
"# Ansible managed",
"#",
"# system_role:ha_cluster",
"",
"",
"",
"## Type: string",
"## Default: \"\"",
"#",
"# SBD_DEVICE specifies the devices to use for exchanging sbd messages",
"# and to monitor. If specifying more than one path, use \";\" as",
"# separator.",
"#",
"SBD_DEVICE=\"\"",
"",
"## Type: yesno",
"## Default: yes",
"#",
"# Whether to enable the pacemaker integration.",
"#",
"SBD_PACEMAKER=yes",
"",
"## Type: always / clean",
"## Default: always",
"#",
"# Specify the start mode for sbd. Setting this to \"clean\" will only",
"# allow sbd to start if it was not previously fenced. See the -S option",
"# in the man page.",
"#",
"SBD_STARTMODE=\"always\"",
"",
"## Type: yesno / integer",
"## Default: no",
"#",
"# Whether to delay after starting sbd on boot for \"msgwait\" seconds.",
"# This may be necessary if your cluster nodes reboot so fast that the",
"# other nodes are still waiting in the fence acknowledgement phase.",
"# This is an occasional issue with virtual machines.",
"#",
"# This can also be enabled by being set to a specific delay value, in",
"# seconds. Sometimes a longer delay than the default, \"msgwait\", is",
"# needed, for example in the cases where it's considered to be safer to",
"# wait longer than:",
"# corosync token timeout + consensus timeout + pcmk_delay_max + msgwait",
"#",
"# Be aware that the special value \"1\" means \"yes\" rather than \"1s\".",
"#",
"# Consider that you might have to adapt the startup-timeout accordingly",
"# if the default isn't sufficient. (TimeoutStartSec for systemd)",
"#",
"# This option may be ignored at a later point, once pacemaker handles",
"# this case better.",
"#",
"SBD_DELAY_START=\"no\"",
"",
"## Type: string",
"## Default: /dev/watchdog",
"#",
"# Watchdog device to use. If set to /dev/null, no watchdog device will",
"# be used.",
"#",
"SBD_WATCHDOG_DEV=\"/dev/watchdog\"",
"",
"## Type: integer",
"## Default: 5",
"#",
"# How long, in seconds, the watchdog will wait before panicking the",
"# node if no-one tickles it.",
"#",
"# This depends mostly on your storage latency; the majority of devices",
"# must be successfully read within this time, or else the node will",
"# self-fence.",
"#",
"# If your sbd device(s) reside on a multipath setup or iSCSI, this",
"# should be the time required to detect a path failure.",
"#",
"# Be aware that watchdog timeout set in the on-disk metadata takes",
"# precedence.",
"#",
"SBD_WATCHDOG_TIMEOUT=\"5\"",
"",
"## Type: string",
"## Default: \"flush,reboot\"",
"#",
"# Actions to be executed when the watchers don't timely report to the sbd",
"# main process or one of the watchers detects that the main process has",
"# died.",
"#",
"# Set timeout-action to comma-separated combination of",
"# noflush|flush plus reboot|crashdump|off.",
"# If just one of both is given the other stays at the default.",
"#",
"# This doesn't affect actions like off, crashdump, reboot explicitly",
"# triggered via message slots.",
"# And it does as well not configure the action a watchdog would",
"# trigger should it run off (there is no generic interface).",
"#",
"SBD_TIMEOUT_ACTION=\"flush,reboot\"",
"",
"## Type: yesno / auto",
"## Default: auto",
"#",
"# If CPUAccounting is enabled default is not to assign any RT-budget",
"# to the system.slice which prevents sbd from running RR-scheduled.",
"#",
"# One way to escape that issue is to move sbd-processes from the",
"# slice they were originally started to root-slice.",
"# Of course starting sbd in a certain slice might be intentional.",
"# Thus in auto-mode sbd will check if the slice has RT-budget assigned.",
"# If that is the case sbd will stay in that slice while it will",
"# be moved to root-slice otherwise.",
"#",
"SBD_MOVE_TO_ROOT_CGROUP=auto",
"",
"## Type: yesno",
"## Default: yes",
"#",
"# If resource startup syncing is enabled then pacemakerd is",
"# gonna wait to be pinged via IPC before it starts resources.",
"# On shutdown pacemakerd is going to wait in a state where it",
"# has cleanly shutdown resources till sbd fetches that state.",
"#",
"# The default is set when building SBD and Pacemaker from source.",
"# Going for 'no' is safer if it can't be assured that SBD and",
"# Pacemaker installed do both support the synchronization feature.",
"# When going with 'yes' - also using package dependencies to",
"# assure SBD & Pacemaker both support the synchronization",
"# feature and are assuming the same default - an SBD configuration",
"# inherited via an upgrade doesn't have to be altered to still",
"# benefit from the new feature.",
"#",
"SBD_SYNC_RESOURCE_STARTUP=yes",
"",
"## Type: string",
"## Default: \"\"",
"#",
"# Additional options for starting sbd",
"#",
"SBD_OPTS=\"-n localhost\""
]
}
TASK [Check SBD config] ********************************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:53
Monday 25 November 2024 18:15:32 -0500 (0:00:00.033) 0:01:22.586 *******
ok: [managed-node2] => {
"changed": false
}
MSG:
All assertions passed
TASK [Fetch quorum configuration] **********************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:61
Monday 25 November 2024 18:15:32 -0500 (0:00:00.036) 0:01:22.623 *******
ok: [managed-node2] => {
"changed": false,
"cmd": [
"pcs",
"quorum",
"config"
],
"delta": "0:00:00.508282",
"end": "2024-11-25 18:15:33.534166",
"rc": 0,
"start": "2024-11-25 18:15:33.025884"
}
STDOUT:
Options:
TASK [Check auto tie breaker in quorum configuration] **************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:67
Monday 25 November 2024 18:15:33 -0500 (0:00:00.906) 0:01:23.530 *******
ok: [managed-node2] => {
"changed": false
}
MSG:
All assertions passed
TASK [Get services status] *****************************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:74
Monday 25 November 2024 18:15:33 -0500 (0:00:00.025) 0:01:23.555 *******
ok: [managed-node2] => {
"ansible_facts": {
"services": {
"NetworkManager-dispatcher.service": {
"name": "NetworkManager-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"NetworkManager-wait-online.service": {
"name": "NetworkManager-wait-online.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"NetworkManager.service": {
"name": "NetworkManager.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"arp-ethers.service": {
"name": "arp-ethers.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"auditd.service": {
"name": "auditd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"auth-rpcgss-module.service": {
"name": "auth-rpcgss-module.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"autovt@.service": {
"name": "autovt@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"blk-availability.service": {
"name": "blk-availability.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"chrony-dnssrv@.service": {
"name": "chrony-dnssrv@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"chrony-wait.service": {
"name": "chrony-wait.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"chronyd.service": {
"name": "chronyd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cloud-config.service": {
"name": "cloud-config.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-final.service": {
"name": "cloud-final.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init-hotplugd.service": {
"name": "cloud-init-hotplugd.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"cloud-init-local.service": {
"name": "cloud-init-local.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"cloud-init.service": {
"name": "cloud-init.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"console-getty.service": {
"name": "console-getty.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"container-getty@.service": {
"name": "container-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"corosync-notifyd.service": {
"name": "corosync-notifyd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qdevice.service": {
"name": "corosync-qdevice.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync-qnetd.service": {
"name": "corosync-qnetd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"corosync.service": {
"name": "corosync.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"cpupower.service": {
"name": "cpupower.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"crm_mon.service": {
"name": "crm_mon.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"crond.service": {
"name": "crond.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"dbus-org.freedesktop.hostname1.service": {
"name": "dbus-org.freedesktop.hostname1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.locale1.service": {
"name": "dbus-org.freedesktop.locale1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.login1.service": {
"name": "dbus-org.freedesktop.login1.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"dbus-org.freedesktop.nm-dispatcher.service": {
"name": "dbus-org.freedesktop.nm-dispatcher.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus-org.freedesktop.portable1.service": {
"name": "dbus-org.freedesktop.portable1.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dbus-org.freedesktop.timedate1.service": {
"name": "dbus-org.freedesktop.timedate1.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"dbus.service": {
"name": "dbus.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"debug-shell.service": {
"name": "debug-shell.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dm-event.service": {
"name": "dm-event.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-makecache.service": {
"name": "dnf-makecache.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dnf-system-upgrade-cleanup.service": {
"name": "dnf-system-upgrade-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"dnf-system-upgrade.service": {
"name": "dnf-system-upgrade.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"dracut-cmdline.service": {
"name": "dracut-cmdline.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-initqueue.service": {
"name": "dracut-initqueue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-mount.service": {
"name": "dracut-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-mount.service": {
"name": "dracut-pre-mount.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-pivot.service": {
"name": "dracut-pre-pivot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-trigger.service": {
"name": "dracut-pre-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-pre-udev.service": {
"name": "dracut-pre-udev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown-onfailure.service": {
"name": "dracut-shutdown-onfailure.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"dracut-shutdown.service": {
"name": "dracut-shutdown.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"ebtables.service": {
"name": "ebtables.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"emergency.service": {
"name": "emergency.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"firewalld.service": {
"name": "firewalld.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"fstrim.service": {
"name": "fstrim.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"getty@.service": {
"name": "getty@.service",
"source": "systemd",
"state": "unknown",
"status": "enabled"
},
"getty@tty1.service": {
"name": "getty@tty1.service",
"source": "systemd",
"state": "running",
"status": "unknown"
},
"grub-boot-indeterminate.service": {
"name": "grub-boot-indeterminate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"gssproxy.service": {
"name": "gssproxy.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"halt-local.service": {
"name": "halt-local.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"import-state.service": {
"name": "import-state.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"initrd-cleanup.service": {
"name": "initrd-cleanup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-parse-etc.service": {
"name": "initrd-parse-etc.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-switch-root.service": {
"name": "initrd-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"initrd-udevadm-cleanup-db.service": {
"name": "initrd-udevadm-cleanup-db.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"iprdump.service": {
"name": "iprdump.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprinit.service": {
"name": "iprinit.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"iprupdate.service": {
"name": "iprupdate.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"irqbalance.service": {
"name": "irqbalance.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"kdump.service": {
"name": "kdump.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"kmod-static-nodes.service": {
"name": "kmod-static-nodes.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"kvm_stat.service": {
"name": "kvm_stat.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"ldconfig.service": {
"name": "ldconfig.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"loadmodules.service": {
"name": "loadmodules.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-lvmpolld.service": {
"name": "lvm2-lvmpolld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"lvm2-monitor.service": {
"name": "lvm2-monitor.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"lvm2-pvscan@.service": {
"name": "lvm2-pvscan@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"man-db-cache-update.service": {
"name": "man-db-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"man-db-restart-cache-update.service": {
"name": "man-db-restart-cache-update.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"messagebus.service": {
"name": "messagebus.service",
"source": "systemd",
"state": "active",
"status": "static"
},
"microcode.service": {
"name": "microcode.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"multipathd.service": {
"name": "multipathd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"nfs-blkmap.service": {
"name": "nfs-blkmap.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-convert.service": {
"name": "nfs-convert.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nfs-idmapd.service": {
"name": "nfs-idmapd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-mountd.service": {
"name": "nfs-mountd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfs-server.service": {
"name": "nfs-server.service",
"source": "systemd",
"state": "stopped",
"status": "disabled"
},
"nfs-utils.service": {
"name": "nfs-utils.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nfsdcld.service": {
"name": "nfsdcld.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"nftables.service": {
"name": "nftables.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"nis-domainname.service": {
"name": "nis-domainname.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"oddjobd.service": {
"name": "oddjobd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"pacemaker.service": {
"name": "pacemaker.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"pcsd-ruby.service": {
"name": "pcsd-ruby.service",
"source": "systemd",
"state": "running",
"status": "disabled"
},
"pcsd.service": {
"name": "pcsd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"plymouth-halt.service": {
"name": "plymouth-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-kexec.service": {
"name": "plymouth-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-poweroff.service": {
"name": "plymouth-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-quit-wait.service": {
"name": "plymouth-quit-wait.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-quit.service": {
"name": "plymouth-quit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-read-write.service": {
"name": "plymouth-read-write.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-reboot.service": {
"name": "plymouth-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-start.service": {
"name": "plymouth-start.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"plymouth-switch-root-initramfs.service": {
"name": "plymouth-switch-root-initramfs.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"plymouth-switch-root.service": {
"name": "plymouth-switch-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"polkit.service": {
"name": "polkit.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"qemu-guest-agent.service": {
"name": "qemu-guest-agent.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"quotaon.service": {
"name": "quotaon.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"rc-local.service": {
"name": "rc-local.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rdisc.service": {
"name": "rdisc.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"rescue.service": {
"name": "rescue.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"restraintd.service": {
"name": "restraintd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rngd.service": {
"name": "rngd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rpc-gssd.service": {
"name": "rpc-gssd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd-notify.service": {
"name": "rpc-statd-notify.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpc-statd.service": {
"name": "rpc-statd.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"rpcbind.service": {
"name": "rpcbind.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"rsyslog.service": {
"name": "rsyslog.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sbd.service": {
"name": "sbd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sbd_remote.service": {
"name": "sbd_remote.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"selinux-autorelabel-mark.service": {
"name": "selinux-autorelabel-mark.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"selinux-autorelabel.service": {
"name": "selinux-autorelabel.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"serial-getty@.service": {
"name": "serial-getty@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@.service": {
"name": "sshd-keygen@.service",
"source": "systemd",
"state": "unknown",
"status": "disabled"
},
"sshd-keygen@ecdsa.service": {
"name": "sshd-keygen@ecdsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@ed25519.service": {
"name": "sshd-keygen@ed25519.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd-keygen@rsa.service": {
"name": "sshd-keygen@rsa.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"sshd.service": {
"name": "sshd.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"sshd@.service": {
"name": "sshd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"sssd-autofs.service": {
"name": "sssd-autofs.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-kcm.service": {
"name": "sssd-kcm.service",
"source": "systemd",
"state": "stopped",
"status": "indirect"
},
"sssd-nss.service": {
"name": "sssd-nss.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pac.service": {
"name": "sssd-pac.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-pam.service": {
"name": "sssd-pam.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-ssh.service": {
"name": "sssd-ssh.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd-sudo.service": {
"name": "sssd-sudo.service",
"source": "systemd",
"state": "inactive",
"status": "indirect"
},
"sssd.service": {
"name": "sssd.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"syslog.service": {
"name": "syslog.service",
"source": "systemd",
"state": "active",
"status": "enabled"
},
"system-update-cleanup.service": {
"name": "system-update-cleanup.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-ask-password-console.service": {
"name": "systemd-ask-password-console.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-plymouth.service": {
"name": "systemd-ask-password-plymouth.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-ask-password-wall.service": {
"name": "systemd-ask-password-wall.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-backlight@.service": {
"name": "systemd-backlight@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-binfmt.service": {
"name": "systemd-binfmt.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-coredump@.service": {
"name": "systemd-coredump@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-exit.service": {
"name": "systemd-exit.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-firstboot.service": {
"name": "systemd-firstboot.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck-root.service": {
"name": "systemd-fsck-root.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-fsck@.service": {
"name": "systemd-fsck@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-halt.service": {
"name": "systemd-halt.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hibernate-resume@.service": {
"name": "systemd-hibernate-resume@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"systemd-hibernate.service": {
"name": "systemd-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hostnamed.service": {
"name": "systemd-hostnamed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-hwdb-update.service": {
"name": "systemd-hwdb-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-hybrid-sleep.service": {
"name": "systemd-hybrid-sleep.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-initctl.service": {
"name": "systemd-initctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-catalog-update.service": {
"name": "systemd-journal-catalog-update.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journal-flush.service": {
"name": "systemd-journal-flush.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-journald.service": {
"name": "systemd-journald.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-kexec.service": {
"name": "systemd-kexec.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-localed.service": {
"name": "systemd-localed.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-logind.service": {
"name": "systemd-logind.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-machine-id-commit.service": {
"name": "systemd-machine-id-commit.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-modules-load.service": {
"name": "systemd-modules-load.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-portabled.service": {
"name": "systemd-portabled.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-poweroff.service": {
"name": "systemd-poweroff.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-pstore.service": {
"name": "systemd-pstore.service",
"source": "systemd",
"state": "stopped",
"status": "enabled"
},
"systemd-quotacheck.service": {
"name": "systemd-quotacheck.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-random-seed.service": {
"name": "systemd-random-seed.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-reboot.service": {
"name": "systemd-reboot.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-remount-fs.service": {
"name": "systemd-remount-fs.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-resolved.service": {
"name": "systemd-resolved.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"systemd-rfkill.service": {
"name": "systemd-rfkill.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend-then-hibernate.service": {
"name": "systemd-suspend-then-hibernate.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-suspend.service": {
"name": "systemd-suspend.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"systemd-sysctl.service": {
"name": "systemd-sysctl.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-sysusers.service": {
"name": "systemd-sysusers.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-timedated.service": {
"name": "systemd-timedated.service",
"source": "systemd",
"state": "inactive",
"status": "masked"
},
"systemd-tmpfiles-clean.service": {
"name": "systemd-tmpfiles-clean.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup-dev.service": {
"name": "systemd-tmpfiles-setup-dev.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-tmpfiles-setup.service": {
"name": "systemd-tmpfiles-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-settle.service": {
"name": "systemd-udev-settle.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udev-trigger.service": {
"name": "systemd-udev-trigger.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-udevd.service": {
"name": "systemd-udevd.service",
"source": "systemd",
"state": "running",
"status": "static"
},
"systemd-update-done.service": {
"name": "systemd-update-done.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp-runlevel.service": {
"name": "systemd-update-utmp-runlevel.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-update-utmp.service": {
"name": "systemd-update-utmp.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-user-sessions.service": {
"name": "systemd-user-sessions.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-vconsole-setup.service": {
"name": "systemd-vconsole-setup.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"systemd-volatile-root.service": {
"name": "systemd-volatile-root.service",
"source": "systemd",
"state": "inactive",
"status": "static"
},
"tcsd.service": {
"name": "tcsd.service",
"source": "systemd",
"state": "inactive",
"status": "disabled"
},
"teamd@.service": {
"name": "teamd@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"timedatex.service": {
"name": "timedatex.service",
"source": "systemd",
"state": "inactive",
"status": "enabled"
},
"tuned.service": {
"name": "tuned.service",
"source": "systemd",
"state": "running",
"status": "enabled"
},
"unbound-anchor.service": {
"name": "unbound-anchor.service",
"source": "systemd",
"state": "stopped",
"status": "static"
},
"user-runtime-dir@.service": {
"name": "user-runtime-dir@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user-runtime-dir@0.service": {
"name": "user-runtime-dir@0.service",
"source": "systemd",
"state": "stopped",
"status": "unknown"
},
"user@.service": {
"name": "user@.service",
"source": "systemd",
"state": "unknown",
"status": "static"
},
"user@0.service": {
"name": "user@0.service",
"source": "systemd",
"state": "running",
"status": "unknown"
}
}
},
"changed": false
}
TASK [Check services status] ***************************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:77
Monday 25 November 2024 18:15:35 -0500 (0:00:01.674) 0:01:25.230 *******
ok: [managed-node2] => {
"changed": false
}
MSG:
All assertions passed
TASK [Check cluster status] ****************************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:84
Monday 25 November 2024 18:15:35 -0500 (0:00:00.029) 0:01:25.260 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/assert_cluster_running.yml for managed-node2
TASK [Get pcs auth status] *****************************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/assert_cluster_running.yml:3
Monday 25 November 2024 18:15:35 -0500 (0:00:00.034) 0:01:25.294 *******
ok: [managed-node2] => {
"changed": false,
"cmd": [
"pcs",
"status",
"pcsd"
],
"delta": "0:00:00.525337",
"end": "2024-11-25 18:15:36.184616",
"rc": 0,
"start": "2024-11-25 18:15:35.659279"
}
STDOUT:
localhost: Online
TASK [Check pcs auth] **********************************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/assert_cluster_running.yml:9
Monday 25 November 2024 18:15:36 -0500 (0:00:00.887) 0:01:26.182 *******
ok: [managed-node2] => {
"changed": false
}
MSG:
All assertions passed
TASK [Get cluster status] ******************************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/assert_cluster_running.yml:14
Monday 25 November 2024 18:15:36 -0500 (0:00:00.028) 0:01:26.211 *******
ok: [managed-node2] => {
"changed": false,
"cmd": [
"pcs",
"status"
],
"delta": "0:00:00.628468",
"end": "2024-11-25 18:15:37.201678",
"rc": 0,
"start": "2024-11-25 18:15:36.573210"
}
STDOUT:
Cluster name: test-cluster
Cluster Summary:
* Stack: corosync (Pacemaker is running)
* Current DC: localhost (version 2.1.7-5.el8-0f7f88312) - partition with quorum
* Last updated: Mon Nov 25 18:15:37 2024 on localhost
* Last change: Mon Nov 25 18:15:29 2024 by root via root on localhost
* 1 node configured
* 0 resource instances configured
Node List:
* Online: [ localhost ]
Full List of Resources:
* No resources
Daemon Status:
corosync: active/enabled
pacemaker: active/enabled
pcsd: active/enabled
sbd: active/enabled
TASK [Check cluster is running] ************************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/assert_cluster_running.yml:20
Monday 25 November 2024 18:15:37 -0500 (0:00:01.002) 0:01:27.213 *******
ok: [managed-node2] => {
"changed": false
}
MSG:
All assertions passed
TASK [Check firewall and selinux state] ****************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:87
Monday 25 November 2024 18:15:37 -0500 (0:00:00.035) 0:01:27.249 *******
included: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/check_firewall_selinux.yml for managed-node2
TASK [Check firewall service status] *******************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/check_firewall_selinux.yml:6
Monday 25 November 2024 18:15:37 -0500 (0:00:00.063) 0:01:27.312 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Check firewall port status] **********************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/check_firewall_selinux.yml:12
Monday 25 November 2024 18:15:37 -0500 (0:00:00.032) 0:01:27.345 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Get associated selinux ports] ********************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/check_firewall_selinux.yml:25
Monday 25 November 2024 18:15:37 -0500 (0:00:00.034) 0:01:27.379 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Check associated selinux ports] ******************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/check_firewall_selinux.yml:33
Monday 25 November 2024 18:15:37 -0500 (0:00:00.035) 0:01:27.415 *******
skipping: [managed-node2] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [Clean up test environment for SBD] ***************************************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:91
Monday 25 November 2024 18:15:37 -0500 (0:00:00.050) 0:01:27.465 *******
TASK [fedora.linux_system_roles.ha_cluster : Unmount SBD devices] **************
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_cleanup_sbd.yml:9
Monday 25 November 2024 18:15:37 -0500 (0:00:00.085) 0:01:27.551 *******
changed: [managed-node2] => {
"changed": true,
"cmd": [
"losetup",
"-d",
"/dev/loop0"
],
"delta": "0:00:00.006607",
"end": "2024-11-25 18:15:37.977522",
"rc": 0,
"start": "2024-11-25 18:15:37.970915"
}
TASK [fedora.linux_system_roles.ha_cluster : Delete backing files for SBD devices] ***
task path: /tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_cleanup_sbd.yml:14
Monday 25 November 2024 18:15:38 -0500 (0:00:00.488) 0:01:28.039 *******
ok: [managed-node2] => {
"changed": false,
"path": "{'changed': True, 'path': '/tmp/ansible.1j7reysj_ha_cluster_tests', 'uid': 0, 'gid': 0, 'owner': 'root', 'group': 'root', 'mode': '0600', 'state': 'file', 'secontext': 'unconfined_u:object_r:user_tmp_t:s0', 'size': 0, 'failed': False}",
"state": "absent"
}
META: ran handlers
META: ran handlers
PLAY RECAP *********************************************************************
managed-node2 : ok=118 changed=28 unreachable=0 failed=0 skipped=85 rescued=0 ignored=0
Monday 25 November 2024 18:15:38 -0500 (0:00:00.400) 0:01:28.440 *******
===============================================================================
fedora.linux_system_roles.ha_cluster : Wait for the cluster to fully start and form membership -- 26.80s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:75
fedora.linux_system_roles.ha_cluster : Install role essential packages --- 3.04s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:11
fedora.linux_system_roles.ha_cluster : Install cluster packages --------- 2.98s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:58
fedora.linux_system_roles.ha_cluster : Stop cluster daemons to reload configuration --- 2.44s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:24
fedora.linux_system_roles.ha_cluster : Populate service facts ----------- 1.95s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/selinux.yml:3
fedora.linux_system_roles.ha_cluster : Get services status - detect pacemaker --- 1.77s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:135
fedora.linux_system_roles.ha_cluster : Get services status - detect corosync-qdevice --- 1.77s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-start-and-reload.yml:21
fedora.linux_system_roles.ha_cluster : Get services status - detect SBD --- 1.76s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_common/cluster-enable-disable.yml:16
Get services status ----------------------------------------------------- 1.67s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:74
fedora.linux_system_roles.ha_cluster : Remove qnetd configuration ------- 1.45s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml:3
fedora.linux_system_roles.ha_cluster : Write CIB configuration ---------- 1.39s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:33
fedora.linux_system_roles.ha_cluster : Start pcsd with updated config files and configure it to start on boot --- 1.29s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:88
Gathering Facts --------------------------------------------------------- 1.17s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_sbd_defaults.yml:9
fedora.linux_system_roles.ha_cluster : Remove CIB tempfiles ------------- 1.10s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/create-and-push-cib.yml:290
fedora.linux_system_roles.ha_cluster : Enable or disable configured cluster services on boot --- 1.08s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_common/cluster-enable-disable.yml:3
fedora.linux_system_roles.ha_cluster : Set stonith-watchdog-timeout cluster property in CIB --- 1.07s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/sbd.yml:142
Get cluster status ------------------------------------------------------ 1.00s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tasks/assert_cluster_running.yml:14
fedora.linux_system_roles.ha_cluster : Create a corosync.conf file content using pcs-0.10 --- 0.96s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-cluster-setup-pcs-0.10.yml:3
fedora.linux_system_roles.ha_cluster : Remove qdevice certificates [CLI] --- 0.94s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-setup-keys.yml:37
fedora.linux_system_roles.ha_cluster : Check pcs auth status ------------ 0.94s
/tmp/collections-Li8/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-auth.yml:3