ansible-playbook [core 2.17.5]
  config file = None
  configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
  ansible python module location = /usr/local/lib/python3.12/site-packages/ansible
  ansible collection location = /tmp/collections-kFV
  executable location = /usr/local/bin/ansible-playbook
  python version = 3.12.6 (main, Sep  9 2024, 00:00:00) [GCC 14.2.1 20240801 (Red Hat 14.2.1-1)] (/usr/bin/python3.12)
  jinja version = 3.1.4
  libyaml = True
No config file found; using defaults
running playbook inside collection fedora.linux_system_roles
redirecting (type: callback) ansible.builtin.debug to ansible.posix.debug
redirecting (type: callback) ansible.builtin.debug to ansible.posix.debug
redirecting (type: callback) ansible.builtin.profile_tasks to ansible.posix.profile_tasks
Skipping callback 'default', as we already have a stdout callback.
Skipping callback 'minimal', as we already have a stdout callback.
Skipping callback 'oneline', as we already have a stdout callback.

PLAYBOOK: tests_cluster_destroy.yml ********************************************
2 plays in /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cluster_destroy.yml

PLAY [all] *********************************************************************

TASK [Include vault variables] *************************************************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cluster_destroy.yml:5
Thursday 10 October 2024  08:25:38 -0400 (0:00:00.008)       0:00:00.008 ****** 
ok: [managed-node3] => {
    "ansible_facts": {
        "ha_cluster_hacluster_password": {
            "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n31303833633366333561656439323930303361333161363239346166656537323933313436\n3432386236656563343237306335323637396239616230353561330a313731623238393238\n62343064666336643930663239383936616465643134646536656532323461356237646133\n3761616633323839633232353637366266350a313163633236376666653238633435306565\n3264623032333736393535663833\n"
        }
    },
    "ansible_included_var_files": [
        "/tmp/ha_cluster-PtG/tests/vars/vault-variables.yml"
    ],
    "changed": false
}

PLAY [Deconfigure cluster] *****************************************************

TASK [Gathering Facts] *********************************************************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cluster_destroy.yml:9
Thursday 10 October 2024  08:25:38 -0400 (0:00:00.018)       0:00:00.027 ****** 
[WARNING]: Platform linux on host managed-node3 is using the discovered Python
interpreter at /usr/bin/python3.12, but future installation of another Python
interpreter could change the meaning of that path. See
https://docs.ansible.com/ansible-
core/2.17/reference_appendices/interpreter_discovery.html for more information.
ok: [managed-node3]

TASK [Set up test environment] *************************************************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cluster_destroy.yml:18
Thursday 10 October 2024  08:25:39 -0400 (0:00:01.188)       0:00:01.216 ****** 
included: fedora.linux_system_roles.ha_cluster for managed-node3

TASK [fedora.linux_system_roles.ha_cluster : Set node name to 'localhost' for single-node clusters] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:9
Thursday 10 October 2024  08:25:39 -0400 (0:00:00.030)       0:00:01.247 ****** 
ok: [managed-node3] => {
    "ansible_facts": {
        "inventory_hostname": "localhost"
    },
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Ensure facts used by tests] *******
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:14
Thursday 10 October 2024  08:25:39 -0400 (0:00:00.039)       0:00:01.287 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "'distribution' not in ansible_facts",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Check if system is ostree] ********
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:22
Thursday 10 October 2024  08:25:39 -0400 (0:00:00.017)       0:00:01.304 ****** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "exists": false
    }
}

TASK [fedora.linux_system_roles.ha_cluster : Set flag to indicate system is ostree] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:27
Thursday 10 October 2024  08:25:39 -0400 (0:00:00.409)       0:00:01.714 ****** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__ha_cluster_is_ostree": false
    },
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Do not try to enable RHEL repositories] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:32
Thursday 10 October 2024  08:25:40 -0400 (0:00:00.023)       0:00:01.738 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ansible_distribution == 'RedHat'",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Copy nss-altfiles ha_cluster users to /etc/passwd] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:41
Thursday 10 October 2024  08:25:40 -0400 (0:00:00.013)       0:00:01.752 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "__ha_cluster_is_ostree | d(false)",
    "skip_reason": "Conditional result was False"
}

TASK [Run HA Cluster role] *****************************************************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cluster_destroy.yml:23
Thursday 10 October 2024  08:25:40 -0400 (0:00:00.034)       0:00:01.786 ****** 
included: fedora.linux_system_roles.ha_cluster for managed-node3

TASK [fedora.linux_system_roles.ha_cluster : Set platform/version specific variables] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:3
Thursday 10 October 2024  08:25:40 -0400 (0:00:00.037)       0:00:01.823 ****** 
included: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.ha_cluster : Ensure ansible_facts used by role] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:2
Thursday 10 October 2024  08:25:40 -0400 (0:00:00.019)       0:00:01.843 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "__ha_cluster_required_facts | difference(ansible_facts.keys() | list) | length > 0",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Check if system is ostree] ********
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:10
Thursday 10 October 2024  08:25:40 -0400 (0:00:00.033)       0:00:01.877 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "not __ha_cluster_is_ostree is defined",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Set flag to indicate system is ostree] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:15
Thursday 10 October 2024  08:25:40 -0400 (0:00:00.017)       0:00:01.895 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "not __ha_cluster_is_ostree is defined",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Set platform/version specific variables] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:19
Thursday 10 October 2024  08:25:40 -0400 (0:00:00.016)       0:00:01.911 ****** 
ok: [managed-node3] => (item=RedHat.yml) => {
    "ansible_facts": {
        "__ha_cluster_cloud_agents_packages": [],
        "__ha_cluster_fence_agent_packages_default": "{{ ['fence-agents-all'] + (['fence-virt'] if ansible_architecture == 'x86_64' else []) }}",
        "__ha_cluster_fullstack_node_packages": [
            "corosync",
            "libknet1-plugins-all",
            "resource-agents",
            "pacemaker"
        ],
        "__ha_cluster_pcs_provider": "pcs-0.10",
        "__ha_cluster_qdevice_node_packages": [
            "corosync-qdevice",
            "bash",
            "coreutils",
            "curl",
            "grep",
            "nss-tools",
            "openssl",
            "sed"
        ],
        "__ha_cluster_repos": [],
        "__ha_cluster_role_essential_packages": [
            "pcs",
            "corosync-qnetd",
            "openssl"
        ],
        "__ha_cluster_sbd_packages": [
            "sbd"
        ],
        "__ha_cluster_services": [
            "corosync",
            "corosync-qdevice",
            "pacemaker"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/RedHat.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "false_condition": "__vars_file is file",
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_10.yml) => {
    "ansible_facts": {
        "__ha_cluster_cloud_agents_packages": [
            "resource-agents-cloud",
            "fence-agents-aliyun",
            "fence-agents-aws",
            "fence-agents-azure-arm",
            "fence-agents-compute",
            "fence-agents-gce",
            "fence-agents-ibm-powervs",
            "fence-agents-ibm-vpc",
            "fence-agents-kubevirt",
            "fence-agents-openstack"
        ],
        "__ha_cluster_repos": [
            {
                "id": "highavailability",
                "name": "HighAvailability"
            },
            {
                "id": "resilientstorage",
                "name": "ResilientStorage"
            }
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/CentOS_10.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_10.yml"
}
ok: [managed-node3] => (item=CentOS_10.yml) => {
    "ansible_facts": {
        "__ha_cluster_cloud_agents_packages": [
            "resource-agents-cloud",
            "fence-agents-aliyun",
            "fence-agents-aws",
            "fence-agents-azure-arm",
            "fence-agents-compute",
            "fence-agents-gce",
            "fence-agents-ibm-powervs",
            "fence-agents-ibm-vpc",
            "fence-agents-kubevirt",
            "fence-agents-openstack"
        ],
        "__ha_cluster_repos": [
            {
                "id": "highavailability",
                "name": "HighAvailability"
            },
            {
                "id": "resilientstorage",
                "name": "ResilientStorage"
            }
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/CentOS_10.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_10.yml"
}

TASK [fedora.linux_system_roles.ha_cluster : Set Linux Pacemaker shell specific variables] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:34
Thursday 10 October 2024  08:25:40 -0400 (0:00:00.041)       0:00:01.953 ****** 
ok: [managed-node3] => {
    "ansible_facts": {},
    "ansible_included_var_files": [
        "/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/shell_pcs.yml"
    ],
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Enable package repositories] ******
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:6
Thursday 10 October 2024  08:25:40 -0400 (0:00:00.016)       0:00:01.970 ****** 
included: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-package-repositories.yml for managed-node3

TASK [fedora.linux_system_roles.ha_cluster : Find platform/version specific tasks to enable repositories] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-package-repositories.yml:3
Thursday 10 October 2024  08:25:40 -0400 (0:00:00.021)       0:00:01.992 ****** 
ok: [managed-node3] => (item=RedHat.yml) => {
    "ansible_facts": {
        "__ha_cluster_enable_repo_tasks_file": "/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/RedHat.yml"
    },
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml"
}
ok: [managed-node3] => (item=CentOS.yml) => {
    "ansible_facts": {
        "__ha_cluster_enable_repo_tasks_file": "/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml"
    },
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml"
}
skipping: [managed-node3] => (item=CentOS_10.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "false_condition": "__ha_cluster_enable_repo_tasks_file_candidate is file",
    "item": "CentOS_10.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS_10.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "false_condition": "__ha_cluster_enable_repo_tasks_file_candidate is file",
    "item": "CentOS_10.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Run platform/version specific tasks to enable repositories] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-package-repositories.yml:21
Thursday 10 October 2024  08:25:40 -0400 (0:00:00.040)       0:00:02.032 ****** 
included: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml for managed-node3

TASK [fedora.linux_system_roles.ha_cluster : List active CentOS repositories] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml:3
Thursday 10 October 2024  08:25:40 -0400 (0:00:00.056)       0:00:02.089 ****** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "dnf",
        "repolist"
    ],
    "delta": "0:00:00.193878",
    "end": "2024-10-10 08:25:40.987106",
    "rc": 0,
    "start": "2024-10-10 08:25:40.793228"
}

STDOUT:

repo id                 repo name
appstream               CentOS Stream 10 - AppStream
baseos                  CentOS Stream 10 - BaseOS
beaker-client           Beaker Client - RedHatEnterpriseLinux9
beaker-harness          Beaker harness
beakerlib-libraries     Copr repo for beakerlib-libraries owned by bgoncalv
highavailability        CentOS Stream 10 - HighAvailability

TASK [fedora.linux_system_roles.ha_cluster : Enable CentOS repositories] *******
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml:10
Thursday 10 October 2024  08:25:41 -0400 (0:00:00.664)       0:00:02.753 ****** 
skipping: [managed-node3] => (item={'id': 'highavailability', 'name': 'HighAvailability'})  => {
    "ansible_loop_var": "item",
    "changed": false,
    "false_condition": "item.id not in __ha_cluster_repolist.stdout",
    "item": {
        "id": "highavailability",
        "name": "HighAvailability"
    },
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item={'id': 'resilientstorage', 'name': 'ResilientStorage'})  => {
    "ansible_loop_var": "item",
    "changed": false,
    "false_condition": "item.name != \"ResilientStorage\" or ha_cluster_enable_repos_resilient_storage",
    "item": {
        "id": "resilientstorage",
        "name": "ResilientStorage"
    },
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => {
    "changed": false
}

MSG:

All items skipped

TASK [fedora.linux_system_roles.ha_cluster : Install role essential packages] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:11
Thursday 10 October 2024  08:25:41 -0400 (0:00:00.025)       0:00:02.779 ****** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.ha_cluster : Check and prepare role variables] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:17
Thursday 10 October 2024  08:25:41 -0400 (0:00:00.790)       0:00:03.569 ****** 
included: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml for managed-node3

TASK [fedora.linux_system_roles.ha_cluster : Discover cluster node names] ******
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:3
Thursday 10 October 2024  08:25:41 -0400 (0:00:00.035)       0:00:03.605 ****** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__ha_cluster_node_name": "localhost"
    },
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Collect cluster node names] *******
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:7
Thursday 10 October 2024  08:25:41 -0400 (0:00:00.041)       0:00:03.646 ****** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__ha_cluster_all_node_names": [
            "localhost"
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Fail if ha_cluster_node_options contains unknown or duplicate nodes] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:16
Thursday 10 October 2024  08:25:41 -0400 (0:00:00.039)       0:00:03.686 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "(\n  __nodes_from_options != (__nodes_from_options | unique)\n) or (\n  __nodes_from_options | difference(__ha_cluster_all_node_names)\n)\n",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Extract node options] *************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:30
Thursday 10 October 2024  08:25:42 -0400 (0:00:00.039)       0:00:03.726 ****** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__ha_cluster_local_node": {}
    },
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Fail if passwords are not specified] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:43
Thursday 10 October 2024  08:25:42 -0400 (0:00:00.050)       0:00:03.776 ****** 
skipping: [managed-node3] => (item=ha_cluster_hacluster_password)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "false_condition": "lookup(\"vars\", item, default=\"\") | string | length < 1",
    "item": "ha_cluster_hacluster_password",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => {
    "changed": false
}

MSG:

All items skipped

TASK [fedora.linux_system_roles.ha_cluster : Fail if nodes do not have the same number of SBD devices specified] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:53
Thursday 10 October 2024  08:25:42 -0400 (0:00:00.075)       0:00:03.852 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_cluster_present",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Fail if configuring qnetd on a cluster node] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:69
Thursday 10 October 2024  08:25:42 -0400 (0:00:00.014)       0:00:03.866 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_cluster_present",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Fail if no valid level is specified for a fencing level] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:79
Thursday 10 October 2024  08:25:42 -0400 (0:00:00.018)       0:00:03.885 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "skipped_reason": "No items in the list"
}

TASK [fedora.linux_system_roles.ha_cluster : Fail if no target is specified for a fencing level] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:87
Thursday 10 October 2024  08:25:42 -0400 (0:00:00.012)       0:00:03.898 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "skipped_reason": "No items in the list"
}

TASK [fedora.linux_system_roles.ha_cluster : Extract qdevice settings] *********
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:101
Thursday 10 October 2024  08:25:42 -0400 (0:00:00.015)       0:00:03.914 ****** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__ha_cluster_qdevice_host": "",
        "__ha_cluster_qdevice_in_use": false,
        "__ha_cluster_qdevice_model": "",
        "__ha_cluster_qdevice_pcs_address": ""
    },
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Figure out if ATB needs to be enabled for SBD] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:110
Thursday 10 October 2024  08:25:42 -0400 (0:00:00.048)       0:00:03.962 ****** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__ha_cluster_sbd_needs_atb": false
    },
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Fail if SBD needs ATB enabled and the user configured ATB to be disabled] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:120
Thursday 10 October 2024  08:25:42 -0400 (0:00:00.037)       0:00:04.000 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "__ha_cluster_sbd_needs_atb | bool",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Fail if ha_cluster_pcsd_public_key_src and ha_cluster_pcsd_private_key_src are set along with ha_cluster_pcsd_certificates] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:127
Thursday 10 October 2024  08:25:42 -0400 (0:00:00.029)       0:00:04.029 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_pcsd_public_key_src is not none",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Fetch pcs capabilities] ***********
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:141
Thursday 10 October 2024  08:25:42 -0400 (0:00:00.018)       0:00:04.047 ****** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "pcs",
        "--version",
        "--full"
    ],
    "delta": "0:00:00.595221",
    "end": "2024-10-10 08:25:43.203512",
    "rc": 0,
    "start": "2024-10-10 08:25:42.608291"
}

STDOUT:

0.12.0a1
booth booth.enable-authfile.unset cluster.config.backup-local cluster.config.restore-cluster cluster.config.restore-local cluster.config.uuid cluster.create cluster.create.enable cluster.create.local cluster.create.no-keys-sync cluster.create.separated-name-and-address cluster.create.start cluster.create.start.wait cluster.create.transport.knet cluster.create.transport.udp-udpu cluster.create.transport.udp-udpu.no-rrp cluster.destroy cluster.destroy.all cluster.report cluster.verify corosync.authkey.update corosync.config.get corosync.config.get.struct corosync.config.reload corosync.config.sync-to-local-cluster corosync.config.update corosync.link.add corosync.link.remove corosync.link.remove.list corosync.link.update corosync.qdevice corosync.qdevice.model.net corosync.quorum corosync.quorum.device corosync.quorum.device.client.model.net.certificates.local corosync.quorum.device.heuristics corosync.quorum.device.model.net corosync.quorum.device.model.net.options_tls_and_kaptb corosync.quorum.set-expected-votes-runtime corosync.quorum.status corosync.quorum.unblock corosync.totem.block_unlisted_ips corosync.uidgid node.add node.add.enable node.add.separated-name-and-address node.add.start node.add.start.wait node.attributes node.attributes.set-list-for-node node.confirm-off node.fence node.guest node.kill node.maintenance node.maintenance.all node.maintenance.list node.maintenance.wait node.remote node.remote.onfail-demote node.remove node.remove-from-caches node.remove.list node.standby node.standby.all node.standby.list node.standby.wait node.start-stop-enable-disable node.start-stop-enable-disable.all node.start-stop-enable-disable.list node.start-stop-enable-disable.start-wait node.utilization node.utilization.set-list-for-node pcmk.acl.enable-disable pcmk.acl.group pcmk.acl.role pcmk.acl.role.create-with-permissions pcmk.acl.role.delete-with-users-groups pcmk.acl.user pcmk.alert pcmk.cib.checkpoints pcmk.cib.checkpoints.diff pcmk.cib.edit pcmk.cib.get pcmk.cib.get.scope pcmk.cib.roles.promoted-unpromoted pcmk.cib.set pcmk.constraint.colocation.set pcmk.constraint.colocation.set.options pcmk.constraint.colocation.simple pcmk.constraint.colocation.simple.options pcmk.constraint.colocation.simple.options.score pcmk.constraint.config.output-formats pcmk.constraint.hide-expired pcmk.constraint.location.simple pcmk.constraint.location.simple.options pcmk.constraint.location.simple.options.score pcmk.constraint.location.simple.resource-regexp pcmk.constraint.location.simple.rule pcmk.constraint.location.simple.rule.node-attr-type-number pcmk.constraint.location.simple.rule.options pcmk.constraint.location.simple.rule.rule-add-remove pcmk.constraint.location.simple.rule.rule-as-one-argument pcmk.constraint.no-autocorrect pcmk.constraint.order.set pcmk.constraint.order.set.options pcmk.constraint.order.simple pcmk.constraint.order.simple.options pcmk.constraint.ticket.set pcmk.constraint.ticket.set.options pcmk.constraint.ticket.simple pcmk.constraint.ticket.simple.constraint-id pcmk.properties.cluster pcmk.properties.cluster.config.output-formats pcmk.properties.cluster.defaults pcmk.properties.cluster.describe pcmk.properties.cluster.describe.output-formats pcmk.properties.cluster.set_properties.iso8601duration pcmk.properties.operation-defaults pcmk.properties.operation-defaults.config.output-formats pcmk.properties.operation-defaults.multiple pcmk.properties.operation-defaults.rule pcmk.properties.operation-defaults.rule-rsc-op pcmk.properties.operation-defaults.rule.hide-expired pcmk.properties.operation-defaults.rule.node-attr-type-number pcmk.properties.resource-defaults pcmk.properties.resource-defaults.config.output-formats pcmk.properties.resource-defaults.multiple pcmk.properties.resource-defaults.rule pcmk.properties.resource-defaults.rule-rsc-op pcmk.properties.resource-defaults.rule.hide-expired pcmk.properties.resource-defaults.rule.node-attr-type-number pcmk.resource.ban-move-clear pcmk.resource.ban-move-clear.bundles pcmk.resource.ban-move-clear.clear-expired pcmk.resource.ban-move-clear.clone pcmk.resource.bundle pcmk.resource.bundle.container-docker pcmk.resource.bundle.container-docker.promoted-max pcmk.resource.bundle.container-podman pcmk.resource.bundle.container-podman.promoted-max pcmk.resource.bundle.container-rkt pcmk.resource.bundle.container-rkt.promoted-max pcmk.resource.bundle.reset pcmk.resource.bundle.wait pcmk.resource.cleanup pcmk.resource.cleanup.one-resource pcmk.resource.cleanup.strict pcmk.resource.clone pcmk.resource.clone.custom-id pcmk.resource.clone.meta-in-create pcmk.resource.clone.wait pcmk.resource.config.output-formats pcmk.resource.create pcmk.resource.create.clone.custom-id pcmk.resource.create.group.future pcmk.resource.create.in-existing-bundle pcmk.resource.create.meta pcmk.resource.create.meta.future pcmk.resource.create.no-master pcmk.resource.create.operations pcmk.resource.create.operations.onfail-demote pcmk.resource.create.promotable pcmk.resource.create.promotable.custom-id pcmk.resource.create.wait pcmk.resource.debug pcmk.resource.delete pcmk.resource.disable.safe pcmk.resource.disable.safe.brief pcmk.resource.disable.safe.tag pcmk.resource.disable.simulate pcmk.resource.disable.simulate.brief pcmk.resource.disable.simulate.tag pcmk.resource.enable-disable pcmk.resource.enable-disable.list pcmk.resource.enable-disable.tag pcmk.resource.enable-disable.wait pcmk.resource.failcount pcmk.resource.group pcmk.resource.group.add-remove-list pcmk.resource.group.wait pcmk.resource.manage-unmanage pcmk.resource.manage-unmanage.list pcmk.resource.manage-unmanage.tag pcmk.resource.manage-unmanage.with-monitor pcmk.resource.move.autoclean pcmk.resource.move.autoclean.default pcmk.resource.promotable pcmk.resource.promotable.custom-id pcmk.resource.promotable.meta-in-create pcmk.resource.promotable.wait pcmk.resource.refresh pcmk.resource.refresh.one-resource pcmk.resource.refresh.strict pcmk.resource.relations pcmk.resource.relocate pcmk.resource.restart pcmk.resource.update pcmk.resource.update-meta pcmk.resource.update-meta.list pcmk.resource.update-meta.wait pcmk.resource.update-operations pcmk.resource.update-operations.onfail-demote pcmk.resource.update.meta pcmk.resource.update.operations pcmk.resource.update.operations.onfail-demote pcmk.resource.update.wait pcmk.resource.utilization pcmk.resource.utilization-set-list-for-resource pcmk.stonith.cleanup pcmk.stonith.cleanup.one-resource pcmk.stonith.cleanup.strict pcmk.stonith.create pcmk.stonith.create.meta pcmk.stonith.create.operations pcmk.stonith.create.operations.onfail-demote pcmk.stonith.create.wait pcmk.stonith.delete pcmk.stonith.enable-disable pcmk.stonith.enable-disable.list pcmk.stonith.enable-disable.wait pcmk.stonith.history.cleanup pcmk.stonith.history.show pcmk.stonith.history.update pcmk.stonith.levels pcmk.stonith.levels.add-remove-devices-list pcmk.stonith.levels.clear pcmk.stonith.levels.node-attr pcmk.stonith.levels.node-regexp pcmk.stonith.levels.verify pcmk.stonith.refresh pcmk.stonith.refresh.one-resource pcmk.stonith.refresh.strict pcmk.stonith.update pcmk.stonith.update.scsi-devices pcmk.stonith.update.scsi-devices.add-remove pcmk.stonith.update.scsi-devices.mpath pcmk.tag pcmk.tag.resources pcs.auth.client pcs.auth.client.cluster pcs.auth.client.token pcs.auth.deauth-client pcs.auth.deauth-server pcs.auth.no-bidirectional pcs.auth.separated-name-and-address pcs.auth.server.token pcs.cfg-in-file.cib pcs.daemon-ssl-cert.set pcs.daemon-ssl-cert.sync-to-local-cluster pcs.disaster-recovery.essentials pcs.reports.severity.deprecation pcs.request-timeout pcs.yes-to-confirm-commands resource-agents.describe resource-agents.list resource-agents.list.detailed resource-agents.ocf.version-1-0 resource-agents.ocf.version-1-1 resource-agents.self-validation sbd sbd.option-timeout-action sbd.shared-block-device status.corosync.membership status.pcmk.query.resource status.pcmk.resources.hide-inactive status.pcmk.resources.id status.pcmk.resources.node status.pcmk.resources.orphaned status.pcmk.wait status.pcmk.xml stonith-agents.describe stonith-agents.list stonith-agents.list.detailed stonith-agents.ocf.version-1-0 stonith-agents.ocf.version-1-1 stonith-agents.self-validation

TASK [fedora.linux_system_roles.ha_cluster : Parse pcs capabilities] ***********
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:148
Thursday 10 October 2024  08:25:43 -0400 (0:00:00.922)       0:00:04.970 ****** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__ha_cluster_pcs_capabilities": [
            "booth",
            "booth.enable-authfile.unset",
            "cluster.config.backup-local",
            "cluster.config.restore-cluster",
            "cluster.config.restore-local",
            "cluster.config.uuid",
            "cluster.create",
            "cluster.create.enable",
            "cluster.create.local",
            "cluster.create.no-keys-sync",
            "cluster.create.separated-name-and-address",
            "cluster.create.start",
            "cluster.create.start.wait",
            "cluster.create.transport.knet",
            "cluster.create.transport.udp-udpu",
            "cluster.create.transport.udp-udpu.no-rrp",
            "cluster.destroy",
            "cluster.destroy.all",
            "cluster.report",
            "cluster.verify",
            "corosync.authkey.update",
            "corosync.config.get",
            "corosync.config.get.struct",
            "corosync.config.reload",
            "corosync.config.sync-to-local-cluster",
            "corosync.config.update",
            "corosync.link.add",
            "corosync.link.remove",
            "corosync.link.remove.list",
            "corosync.link.update",
            "corosync.qdevice",
            "corosync.qdevice.model.net",
            "corosync.quorum",
            "corosync.quorum.device",
            "corosync.quorum.device.client.model.net.certificates.local",
            "corosync.quorum.device.heuristics",
            "corosync.quorum.device.model.net",
            "corosync.quorum.device.model.net.options_tls_and_kaptb",
            "corosync.quorum.set-expected-votes-runtime",
            "corosync.quorum.status",
            "corosync.quorum.unblock",
            "corosync.totem.block_unlisted_ips",
            "corosync.uidgid",
            "node.add",
            "node.add.enable",
            "node.add.separated-name-and-address",
            "node.add.start",
            "node.add.start.wait",
            "node.attributes",
            "node.attributes.set-list-for-node",
            "node.confirm-off",
            "node.fence",
            "node.guest",
            "node.kill",
            "node.maintenance",
            "node.maintenance.all",
            "node.maintenance.list",
            "node.maintenance.wait",
            "node.remote",
            "node.remote.onfail-demote",
            "node.remove",
            "node.remove-from-caches",
            "node.remove.list",
            "node.standby",
            "node.standby.all",
            "node.standby.list",
            "node.standby.wait",
            "node.start-stop-enable-disable",
            "node.start-stop-enable-disable.all",
            "node.start-stop-enable-disable.list",
            "node.start-stop-enable-disable.start-wait",
            "node.utilization",
            "node.utilization.set-list-for-node",
            "pcmk.acl.enable-disable",
            "pcmk.acl.group",
            "pcmk.acl.role",
            "pcmk.acl.role.create-with-permissions",
            "pcmk.acl.role.delete-with-users-groups",
            "pcmk.acl.user",
            "pcmk.alert",
            "pcmk.cib.checkpoints",
            "pcmk.cib.checkpoints.diff",
            "pcmk.cib.edit",
            "pcmk.cib.get",
            "pcmk.cib.get.scope",
            "pcmk.cib.roles.promoted-unpromoted",
            "pcmk.cib.set",
            "pcmk.constraint.colocation.set",
            "pcmk.constraint.colocation.set.options",
            "pcmk.constraint.colocation.simple",
            "pcmk.constraint.colocation.simple.options",
            "pcmk.constraint.colocation.simple.options.score",
            "pcmk.constraint.config.output-formats",
            "pcmk.constraint.hide-expired",
            "pcmk.constraint.location.simple",
            "pcmk.constraint.location.simple.options",
            "pcmk.constraint.location.simple.options.score",
            "pcmk.constraint.location.simple.resource-regexp",
            "pcmk.constraint.location.simple.rule",
            "pcmk.constraint.location.simple.rule.node-attr-type-number",
            "pcmk.constraint.location.simple.rule.options",
            "pcmk.constraint.location.simple.rule.rule-add-remove",
            "pcmk.constraint.location.simple.rule.rule-as-one-argument",
            "pcmk.constraint.no-autocorrect",
            "pcmk.constraint.order.set",
            "pcmk.constraint.order.set.options",
            "pcmk.constraint.order.simple",
            "pcmk.constraint.order.simple.options",
            "pcmk.constraint.ticket.set",
            "pcmk.constraint.ticket.set.options",
            "pcmk.constraint.ticket.simple",
            "pcmk.constraint.ticket.simple.constraint-id",
            "pcmk.properties.cluster",
            "pcmk.properties.cluster.config.output-formats",
            "pcmk.properties.cluster.defaults",
            "pcmk.properties.cluster.describe",
            "pcmk.properties.cluster.describe.output-formats",
            "pcmk.properties.cluster.set_properties.iso8601duration",
            "pcmk.properties.operation-defaults",
            "pcmk.properties.operation-defaults.config.output-formats",
            "pcmk.properties.operation-defaults.multiple",
            "pcmk.properties.operation-defaults.rule",
            "pcmk.properties.operation-defaults.rule-rsc-op",
            "pcmk.properties.operation-defaults.rule.hide-expired",
            "pcmk.properties.operation-defaults.rule.node-attr-type-number",
            "pcmk.properties.resource-defaults",
            "pcmk.properties.resource-defaults.config.output-formats",
            "pcmk.properties.resource-defaults.multiple",
            "pcmk.properties.resource-defaults.rule",
            "pcmk.properties.resource-defaults.rule-rsc-op",
            "pcmk.properties.resource-defaults.rule.hide-expired",
            "pcmk.properties.resource-defaults.rule.node-attr-type-number",
            "pcmk.resource.ban-move-clear",
            "pcmk.resource.ban-move-clear.bundles",
            "pcmk.resource.ban-move-clear.clear-expired",
            "pcmk.resource.ban-move-clear.clone",
            "pcmk.resource.bundle",
            "pcmk.resource.bundle.container-docker",
            "pcmk.resource.bundle.container-docker.promoted-max",
            "pcmk.resource.bundle.container-podman",
            "pcmk.resource.bundle.container-podman.promoted-max",
            "pcmk.resource.bundle.container-rkt",
            "pcmk.resource.bundle.container-rkt.promoted-max",
            "pcmk.resource.bundle.reset",
            "pcmk.resource.bundle.wait",
            "pcmk.resource.cleanup",
            "pcmk.resource.cleanup.one-resource",
            "pcmk.resource.cleanup.strict",
            "pcmk.resource.clone",
            "pcmk.resource.clone.custom-id",
            "pcmk.resource.clone.meta-in-create",
            "pcmk.resource.clone.wait",
            "pcmk.resource.config.output-formats",
            "pcmk.resource.create",
            "pcmk.resource.create.clone.custom-id",
            "pcmk.resource.create.group.future",
            "pcmk.resource.create.in-existing-bundle",
            "pcmk.resource.create.meta",
            "pcmk.resource.create.meta.future",
            "pcmk.resource.create.no-master",
            "pcmk.resource.create.operations",
            "pcmk.resource.create.operations.onfail-demote",
            "pcmk.resource.create.promotable",
            "pcmk.resource.create.promotable.custom-id",
            "pcmk.resource.create.wait",
            "pcmk.resource.debug",
            "pcmk.resource.delete",
            "pcmk.resource.disable.safe",
            "pcmk.resource.disable.safe.brief",
            "pcmk.resource.disable.safe.tag",
            "pcmk.resource.disable.simulate",
            "pcmk.resource.disable.simulate.brief",
            "pcmk.resource.disable.simulate.tag",
            "pcmk.resource.enable-disable",
            "pcmk.resource.enable-disable.list",
            "pcmk.resource.enable-disable.tag",
            "pcmk.resource.enable-disable.wait",
            "pcmk.resource.failcount",
            "pcmk.resource.group",
            "pcmk.resource.group.add-remove-list",
            "pcmk.resource.group.wait",
            "pcmk.resource.manage-unmanage",
            "pcmk.resource.manage-unmanage.list",
            "pcmk.resource.manage-unmanage.tag",
            "pcmk.resource.manage-unmanage.with-monitor",
            "pcmk.resource.move.autoclean",
            "pcmk.resource.move.autoclean.default",
            "pcmk.resource.promotable",
            "pcmk.resource.promotable.custom-id",
            "pcmk.resource.promotable.meta-in-create",
            "pcmk.resource.promotable.wait",
            "pcmk.resource.refresh",
            "pcmk.resource.refresh.one-resource",
            "pcmk.resource.refresh.strict",
            "pcmk.resource.relations",
            "pcmk.resource.relocate",
            "pcmk.resource.restart",
            "pcmk.resource.update",
            "pcmk.resource.update-meta",
            "pcmk.resource.update-meta.list",
            "pcmk.resource.update-meta.wait",
            "pcmk.resource.update-operations",
            "pcmk.resource.update-operations.onfail-demote",
            "pcmk.resource.update.meta",
            "pcmk.resource.update.operations",
            "pcmk.resource.update.operations.onfail-demote",
            "pcmk.resource.update.wait",
            "pcmk.resource.utilization",
            "pcmk.resource.utilization-set-list-for-resource",
            "pcmk.stonith.cleanup",
            "pcmk.stonith.cleanup.one-resource",
            "pcmk.stonith.cleanup.strict",
            "pcmk.stonith.create",
            "pcmk.stonith.create.meta",
            "pcmk.stonith.create.operations",
            "pcmk.stonith.create.operations.onfail-demote",
            "pcmk.stonith.create.wait",
            "pcmk.stonith.delete",
            "pcmk.stonith.enable-disable",
            "pcmk.stonith.enable-disable.list",
            "pcmk.stonith.enable-disable.wait",
            "pcmk.stonith.history.cleanup",
            "pcmk.stonith.history.show",
            "pcmk.stonith.history.update",
            "pcmk.stonith.levels",
            "pcmk.stonith.levels.add-remove-devices-list",
            "pcmk.stonith.levels.clear",
            "pcmk.stonith.levels.node-attr",
            "pcmk.stonith.levels.node-regexp",
            "pcmk.stonith.levels.verify",
            "pcmk.stonith.refresh",
            "pcmk.stonith.refresh.one-resource",
            "pcmk.stonith.refresh.strict",
            "pcmk.stonith.update",
            "pcmk.stonith.update.scsi-devices",
            "pcmk.stonith.update.scsi-devices.add-remove",
            "pcmk.stonith.update.scsi-devices.mpath",
            "pcmk.tag",
            "pcmk.tag.resources",
            "pcs.auth.client",
            "pcs.auth.client.cluster",
            "pcs.auth.client.token",
            "pcs.auth.deauth-client",
            "pcs.auth.deauth-server",
            "pcs.auth.no-bidirectional",
            "pcs.auth.separated-name-and-address",
            "pcs.auth.server.token",
            "pcs.cfg-in-file.cib",
            "pcs.daemon-ssl-cert.set",
            "pcs.daemon-ssl-cert.sync-to-local-cluster",
            "pcs.disaster-recovery.essentials",
            "pcs.reports.severity.deprecation",
            "pcs.request-timeout",
            "pcs.yes-to-confirm-commands",
            "resource-agents.describe",
            "resource-agents.list",
            "resource-agents.list.detailed",
            "resource-agents.ocf.version-1-0",
            "resource-agents.ocf.version-1-1",
            "resource-agents.self-validation",
            "sbd",
            "sbd.option-timeout-action",
            "sbd.shared-block-device",
            "status.corosync.membership",
            "status.pcmk.query.resource",
            "status.pcmk.resources.hide-inactive",
            "status.pcmk.resources.id",
            "status.pcmk.resources.node",
            "status.pcmk.resources.orphaned",
            "status.pcmk.wait",
            "status.pcmk.xml",
            "stonith-agents.describe",
            "stonith-agents.list",
            "stonith-agents.list.detailed",
            "stonith-agents.ocf.version-1-0",
            "stonith-agents.ocf.version-1-1",
            "stonith-agents.self-validation"
        ],
        "__ha_cluster_pcsd_capabilities_available": true
    },
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Fetch pcsd capabilities] **********
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:155
Thursday 10 October 2024  08:25:43 -0400 (0:00:00.026)       0:00:04.996 ****** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "pcsd",
        "--version",
        "--full"
    ],
    "delta": "0:00:00.650239",
    "end": "2024-10-10 08:25:44.213117",
    "rc": 0,
    "start": "2024-10-10 08:25:43.562878"
}

STDOUT:

0.12.0a1
booth.get-config booth.set-config booth.set-config.multiple cluster.config.restore-local cluster.config.uuid cluster.create cluster.create.enable cluster.create.no-keys-sync cluster.create.separated-name-and-address cluster.create.start cluster.create.start.wait cluster.create.transport.knet cluster.create.transport.udp-udpu cluster.create.transport.udp-udpu.no-rrp cluster.destroy cluster.destroy.all corosync.config.get corosync.config.set corosync.qdevice.model.net.certificates corosync.quorum.device.client corosync.quorum.device.client.model.net.certificates corosync.quorum.device.client.model.net.certificates.local corosync.quorum.device.client.model.net.certificates.rest-api.v2 corosync.quorum.status corosync.totem.block_unlisted_ips node.add node.add.enable node.add.list node.add.separated-name-and-address node.add.start node.add.start.wait node.attributes node.maintenance node.maintenance.all node.maintenance.list node.maintenance.wait node.remove node.remove-from-caches node.remove.list node.standby node.standby.all node.standby.list node.standby.wait node.start-stop-enable-disable node.start-stop-enable-disable.all node.start-stop-enable-disable.stop-component node.utilization pcmk.acl.group pcmk.acl.role pcmk.acl.role.create-with-permissions pcmk.acl.role.delete-with-users-groups pcmk.acl.role.delete-with-users-groups-implicit pcmk.acl.user pcmk.alert pcmk.cib.get pcmk.cib.remove_elements.constraints pcmk.cib.remove_elements.rest-api.v2 pcmk.cib.roles.promoted-unpromoted pcmk.constraint.colocation.set pcmk.constraint.colocation.set.options pcmk.constraint.colocation.simple pcmk.constraint.config.output-formats pcmk.constraint.location.simple pcmk.constraint.location.simple.rule pcmk.constraint.location.simple.rule.node-attr-type-number pcmk.constraint.location.simple.rule.rest-api.v2 pcmk.constraint.no-autocorrect pcmk.constraint.order.set pcmk.constraint.order.set.options pcmk.constraint.order.simple pcmk.constraint.ticket.set pcmk.constraint.ticket.set.options pcmk.constraint.ticket.simple pcmk.constraint.ticket.simple.constraint-id pcmk.properties.cluster pcmk.properties.cluster.describe pcmk.properties.cluster.describe.rest-api.v2 pcmk.properties.cluster.get_properties.rest-api.v2 pcmk.properties.cluster.set_properties.iso8601duration pcmk.properties.cluster.set_properties.rest-api.v2 pcmk.properties.operation-defaults.config.rest-api.v2 pcmk.properties.resource-defaults.config.rest-api.v2 pcmk.resource.ban-move-clear pcmk.resource.ban-move-clear.bundles pcmk.resource.ban-move-clear.clear-expired pcmk.resource.ban-move-clear.clone pcmk.resource.cleanup.one-resource pcmk.resource.cleanup.strict pcmk.resource.clone pcmk.resource.create pcmk.resource.create.clone.custom-id pcmk.resource.create.meta pcmk.resource.create.no-master pcmk.resource.create.operations pcmk.resource.create.operations.onfail-demote pcmk.resource.create.promotable pcmk.resource.create.promotable.custom-id pcmk.resource.create.wait pcmk.resource.delete pcmk.resource.delete.list pcmk.resource.disable.safe pcmk.resource.disable.simulate pcmk.resource.disable.simulate.brief pcmk.resource.disable.simulate.tag pcmk.resource.enable-disable pcmk.resource.enable-disable.list pcmk.resource.enable-disable.wait pcmk.resource.group pcmk.resource.manage-unmanage pcmk.resource.manage-unmanage.list pcmk.resource.manage-unmanage.tag pcmk.resource.manage-unmanage.with-monitor pcmk.resource.move.autoclean pcmk.resource.promotable pcmk.resource.refresh.one-resource pcmk.resource.refresh.strict pcmk.resource.update pcmk.resource.update-meta pcmk.resource.utilization pcmk.stonith.cleanup.one-resource pcmk.stonith.cleanup.strict pcmk.stonith.create pcmk.stonith.create.meta pcmk.stonith.create.operations pcmk.stonith.create.operations.onfail-demote pcmk.stonith.create.wait pcmk.stonith.delete pcmk.stonith.delete.list pcmk.stonith.enable-disable pcmk.stonith.enable-disable.list pcmk.stonith.enable-disable.wait pcmk.stonith.levels pcmk.stonith.levels.add-remove-devices-list pcmk.stonith.levels.clear pcmk.stonith.levels.node-attr pcmk.stonith.levels.node-regexp pcmk.stonith.levels.verify pcmk.stonith.refresh.one-resource pcmk.stonith.refresh.strict pcmk.stonith.scsi-unfence-node-mpath pcmk.stonith.scsi-unfence-node-v2 pcmk.stonith.update pcs.auth.export-cluster-known-hosts pcs.auth.known-host-change pcs.auth.no-bidirectional pcs.auth.separated-name-and-address pcs.auth.server pcs.automatic-pcs-configs-sync pcs.daemon-ssl-cert.set pcs.permissions pcs.rest-api.v1.1 pcs.rest-api.v2 resource-agents.describe resource-agents.list resource-agents.list.detailed resource-agents.ocf.version-1-0 resource-agents.ocf.version-1-1 resource-agents.self-validation resource-agents.simplified-query-api sbd sbd-node sbd-node.shared-block-device sbd.option-timeout-action status.pcmk.local-node status.pcmk.resources.rest-api.v2 status.pcmk.wait stonith-agents.describe stonith-agents.list stonith-agents.list.detailed stonith-agents.ocf.version-1-0 stonith-agents.ocf.version-1-1 stonith-agents.self-validation

TASK [fedora.linux_system_roles.ha_cluster : Parse pcsd capabilities] **********
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:163
Thursday 10 October 2024  08:25:44 -0400 (0:00:00.994)       0:00:05.990 ****** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__ha_cluster_pcsd_capabilities": [
            "booth.get-config",
            "booth.set-config",
            "booth.set-config.multiple",
            "cluster.config.restore-local",
            "cluster.config.uuid",
            "cluster.create",
            "cluster.create.enable",
            "cluster.create.no-keys-sync",
            "cluster.create.separated-name-and-address",
            "cluster.create.start",
            "cluster.create.start.wait",
            "cluster.create.transport.knet",
            "cluster.create.transport.udp-udpu",
            "cluster.create.transport.udp-udpu.no-rrp",
            "cluster.destroy",
            "cluster.destroy.all",
            "corosync.config.get",
            "corosync.config.set",
            "corosync.qdevice.model.net.certificates",
            "corosync.quorum.device.client",
            "corosync.quorum.device.client.model.net.certificates",
            "corosync.quorum.device.client.model.net.certificates.local",
            "corosync.quorum.device.client.model.net.certificates.rest-api.v2",
            "corosync.quorum.status",
            "corosync.totem.block_unlisted_ips",
            "node.add",
            "node.add.enable",
            "node.add.list",
            "node.add.separated-name-and-address",
            "node.add.start",
            "node.add.start.wait",
            "node.attributes",
            "node.maintenance",
            "node.maintenance.all",
            "node.maintenance.list",
            "node.maintenance.wait",
            "node.remove",
            "node.remove-from-caches",
            "node.remove.list",
            "node.standby",
            "node.standby.all",
            "node.standby.list",
            "node.standby.wait",
            "node.start-stop-enable-disable",
            "node.start-stop-enable-disable.all",
            "node.start-stop-enable-disable.stop-component",
            "node.utilization",
            "pcmk.acl.group",
            "pcmk.acl.role",
            "pcmk.acl.role.create-with-permissions",
            "pcmk.acl.role.delete-with-users-groups",
            "pcmk.acl.role.delete-with-users-groups-implicit",
            "pcmk.acl.user",
            "pcmk.alert",
            "pcmk.cib.get",
            "pcmk.cib.remove_elements.constraints",
            "pcmk.cib.remove_elements.rest-api.v2",
            "pcmk.cib.roles.promoted-unpromoted",
            "pcmk.constraint.colocation.set",
            "pcmk.constraint.colocation.set.options",
            "pcmk.constraint.colocation.simple",
            "pcmk.constraint.config.output-formats",
            "pcmk.constraint.location.simple",
            "pcmk.constraint.location.simple.rule",
            "pcmk.constraint.location.simple.rule.node-attr-type-number",
            "pcmk.constraint.location.simple.rule.rest-api.v2",
            "pcmk.constraint.no-autocorrect",
            "pcmk.constraint.order.set",
            "pcmk.constraint.order.set.options",
            "pcmk.constraint.order.simple",
            "pcmk.constraint.ticket.set",
            "pcmk.constraint.ticket.set.options",
            "pcmk.constraint.ticket.simple",
            "pcmk.constraint.ticket.simple.constraint-id",
            "pcmk.properties.cluster",
            "pcmk.properties.cluster.describe",
            "pcmk.properties.cluster.describe.rest-api.v2",
            "pcmk.properties.cluster.get_properties.rest-api.v2",
            "pcmk.properties.cluster.set_properties.iso8601duration",
            "pcmk.properties.cluster.set_properties.rest-api.v2",
            "pcmk.properties.operation-defaults.config.rest-api.v2",
            "pcmk.properties.resource-defaults.config.rest-api.v2",
            "pcmk.resource.ban-move-clear",
            "pcmk.resource.ban-move-clear.bundles",
            "pcmk.resource.ban-move-clear.clear-expired",
            "pcmk.resource.ban-move-clear.clone",
            "pcmk.resource.cleanup.one-resource",
            "pcmk.resource.cleanup.strict",
            "pcmk.resource.clone",
            "pcmk.resource.create",
            "pcmk.resource.create.clone.custom-id",
            "pcmk.resource.create.meta",
            "pcmk.resource.create.no-master",
            "pcmk.resource.create.operations",
            "pcmk.resource.create.operations.onfail-demote",
            "pcmk.resource.create.promotable",
            "pcmk.resource.create.promotable.custom-id",
            "pcmk.resource.create.wait",
            "pcmk.resource.delete",
            "pcmk.resource.delete.list",
            "pcmk.resource.disable.safe",
            "pcmk.resource.disable.simulate",
            "pcmk.resource.disable.simulate.brief",
            "pcmk.resource.disable.simulate.tag",
            "pcmk.resource.enable-disable",
            "pcmk.resource.enable-disable.list",
            "pcmk.resource.enable-disable.wait",
            "pcmk.resource.group",
            "pcmk.resource.manage-unmanage",
            "pcmk.resource.manage-unmanage.list",
            "pcmk.resource.manage-unmanage.tag",
            "pcmk.resource.manage-unmanage.with-monitor",
            "pcmk.resource.move.autoclean",
            "pcmk.resource.promotable",
            "pcmk.resource.refresh.one-resource",
            "pcmk.resource.refresh.strict",
            "pcmk.resource.update",
            "pcmk.resource.update-meta",
            "pcmk.resource.utilization",
            "pcmk.stonith.cleanup.one-resource",
            "pcmk.stonith.cleanup.strict",
            "pcmk.stonith.create",
            "pcmk.stonith.create.meta",
            "pcmk.stonith.create.operations",
            "pcmk.stonith.create.operations.onfail-demote",
            "pcmk.stonith.create.wait",
            "pcmk.stonith.delete",
            "pcmk.stonith.delete.list",
            "pcmk.stonith.enable-disable",
            "pcmk.stonith.enable-disable.list",
            "pcmk.stonith.enable-disable.wait",
            "pcmk.stonith.levels",
            "pcmk.stonith.levels.add-remove-devices-list",
            "pcmk.stonith.levels.clear",
            "pcmk.stonith.levels.node-attr",
            "pcmk.stonith.levels.node-regexp",
            "pcmk.stonith.levels.verify",
            "pcmk.stonith.refresh.one-resource",
            "pcmk.stonith.refresh.strict",
            "pcmk.stonith.scsi-unfence-node-mpath",
            "pcmk.stonith.scsi-unfence-node-v2",
            "pcmk.stonith.update",
            "pcs.auth.export-cluster-known-hosts",
            "pcs.auth.known-host-change",
            "pcs.auth.no-bidirectional",
            "pcs.auth.separated-name-and-address",
            "pcs.auth.server",
            "pcs.automatic-pcs-configs-sync",
            "pcs.daemon-ssl-cert.set",
            "pcs.permissions",
            "pcs.rest-api.v1.1",
            "pcs.rest-api.v2",
            "resource-agents.describe",
            "resource-agents.list",
            "resource-agents.list.detailed",
            "resource-agents.ocf.version-1-0",
            "resource-agents.ocf.version-1-1",
            "resource-agents.self-validation",
            "resource-agents.simplified-query-api",
            "sbd",
            "sbd-node",
            "sbd-node.shared-block-device",
            "sbd.option-timeout-action",
            "status.pcmk.local-node",
            "status.pcmk.resources.rest-api.v2",
            "status.pcmk.wait",
            "stonith-agents.describe",
            "stonith-agents.list",
            "stonith-agents.list.detailed",
            "stonith-agents.ocf.version-1-0",
            "stonith-agents.ocf.version-1-1",
            "stonith-agents.self-validation"
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Fail if pcs is to old to configure resources and operations defaults] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:172
Thursday 10 October 2024  08:25:44 -0400 (0:00:00.061)       0:00:06.052 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "( ha_cluster_resource_defaults and not 'pcmk.properties.resource-defaults.multiple' in __ha_cluster_pcs_capabilities ) or ( ha_cluster_resource_operation_defaults and not 'pcmk.properties.operation-defaults.multiple' in __ha_cluster_pcs_capabilities )",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Generate a password hash] *********
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:26
Thursday 10 October 2024  08:25:44 -0400 (0:00:00.025)       0:00:06.077 ****** 
ok: [managed-node3] => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Set hacluster password] ***********
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:38
Thursday 10 October 2024  08:25:44 -0400 (0:00:00.450)       0:00:06.528 ****** 
ok: [managed-node3] => {
    "append": false,
    "changed": false,
    "comment": "cluster user",
    "group": 189,
    "home": "/var/lib/pacemaker",
    "move_home": false,
    "name": "hacluster",
    "password": "NOT_LOGGING_PASSWORD",
    "shell": "/sbin/nologin",
    "state": "present",
    "uid": 189
}

TASK [fedora.linux_system_roles.ha_cluster : Configure shell] ******************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:43
Thursday 10 October 2024  08:25:45 -0400 (0:00:00.582)       0:00:07.111 ****** 
included: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml for managed-node3

TASK [fedora.linux_system_roles.ha_cluster : Stop pcsd] ************************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:6
Thursday 10 October 2024  08:25:45 -0400 (0:00:00.032)       0:00:07.144 ****** 
changed: [managed-node3] => {
    "changed": true,
    "name": "pcsd",
    "state": "stopped",
    "status": {
        "AccessSELinuxContext": "system_u:object_r:cluster_unit_file_t:s0",
        "ActiveEnterTimestamp": "Thu 2024-10-10 08:24:25 EDT",
        "ActiveEnterTimestampMonotonic": "1207512227",
        "ActiveExitTimestamp": "Thu 2024-10-10 08:24:22 EDT",
        "ActiveExitTimestampMonotonic": "1204693235",
        "ActiveState": "active",
        "After": "systemd-journald.socket network-online.target sysinit.target pcsd-ruby.service basic.target system.slice",
        "AllowIsolate": "no",
        "AssertResult": "yes",
        "AssertTimestamp": "Thu 2024-10-10 08:24:24 EDT",
        "AssertTimestampMonotonic": "1206731333",
        "Before": "multi-user.target shutdown.target",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "yes",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "9508455000",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "yes",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore",
        "CleanResult": "success",
        "CollectMode": "inactive",
        "ConditionResult": "yes",
        "ConditionTimestamp": "Thu 2024-10-10 08:24:24 EDT",
        "ConditionTimestampMonotonic": "1206731330",
        "ConfigurationDirectoryMode": "0755",
        "Conflicts": "shutdown.target",
        "ConsistsOf": "pcsd-ruby.service",
        "ControlGroup": "/system.slice/pcsd.service",
        "ControlGroupId": "9010",
        "ControlPID": "0",
        "CoredumpFilter": "0x33",
        "CoredumpReceive": "no",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "DefaultStartupMemoryLow": "0",
        "Delegate": "no",
        "Description": "PCS GUI and remote configuration interface",
        "DevicePolicy": "auto",
        "Documentation": "\"man:pcsd(8)\" \"man:pcs(8)\"",
        "DynamicUser": "no",
        "EffectiveMemoryHigh": "3702874112",
        "EffectiveMemoryMax": "3702874112",
        "EffectiveTasksMax": "22362",
        "EnvironmentFiles": "/etc/sysconfig/pcsd (ignore_errors=no)",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainHandoffTimestamp": "Thu 2024-10-10 08:24:24 EDT",
        "ExecMainHandoffTimestampMonotonic": "1206745416",
        "ExecMainPID": "114677",
        "ExecMainStartTimestamp": "Thu 2024-10-10 08:24:24 EDT",
        "ExecMainStartTimestampMonotonic": "1206732638",
        "ExecMainStatus": "0",
        "ExecStart": "{ path=/usr/sbin/pcsd ; argv[]=/usr/sbin/pcsd ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
        "ExecStartEx": "{ path=/usr/sbin/pcsd ; argv[]=/usr/sbin/pcsd ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
        "ExitType": "main",
        "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FileDescriptorStorePreserve": "restart",
        "FinalKillSignal": "9",
        "FragmentPath": "/usr/lib/systemd/system/pcsd.service",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOReadBytes": "[not set]",
        "IOReadOperations": "[not set]",
        "IOSchedulingClass": "2",
        "IOSchedulingPriority": "4",
        "IOWeight": "[not set]",
        "IOWriteBytes": "[not set]",
        "IOWriteOperations": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "[no data]",
        "IPEgressPackets": "[no data]",
        "IPIngressBytes": "[no data]",
        "IPIngressPackets": "[no data]",
        "Id": "pcsd.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestamp": "Thu 2024-10-10 08:24:22 EDT",
        "InactiveEnterTimestampMonotonic": "1204803114",
        "InactiveExitTimestamp": "Thu 2024-10-10 08:24:24 EDT",
        "InactiveExitTimestampMonotonic": "1206734418",
        "InvocationID": "aa68975f714248cdbf40efb72dc43515",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "mixed",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "8388608",
        "LimitMEMLOCKSoft": "8388608",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "524288",
        "LimitNOFILESoft": "1024",
        "LimitNPROC": "13976",
        "LimitNPROCSoft": "13976",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "13976",
        "LimitSIGPENDINGSoft": "13976",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadState": "loaded",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "114677",
        "ManagedOOMMemoryPressure": "auto",
        "ManagedOOMMemoryPressureLimit": "0",
        "ManagedOOMPreference": "none",
        "ManagedOOMSwap": "auto",
        "MemoryAccounting": "yes",
        "MemoryAvailable": "2661621760",
        "MemoryCurrent": "412516352",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryKSM": "no",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemoryPeak": "414593024",
        "MemoryPressureThresholdUSec": "200ms",
        "MemoryPressureWatch": "auto",
        "MemorySwapCurrent": "0",
        "MemorySwapMax": "infinity",
        "MemorySwapPeak": "0",
        "MemoryZSwapCurrent": "0",
        "MemoryZSwapMax": "infinity",
        "MemoryZSwapWriteback": "yes",
        "MountAPIVFS": "no",
        "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAPolicy": "n/a",
        "Names": "pcsd.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "main",
        "OOMPolicy": "stop",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "OnSuccessJobMode": "fail",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateIPC": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProcSubset": "all",
        "ProtectClock": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectHostname": "no",
        "ProtectKernelLogs": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectProc": "default",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "ReloadResult": "success",
        "ReloadSignal": "1",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Requires": "network-online.target sysinit.target pcsd-ruby.service system.slice",
        "Restart": "no",
        "RestartKillSignal": "15",
        "RestartMaxDelayUSec": "infinity",
        "RestartMode": "normal",
        "RestartSteps": "0",
        "RestartUSec": "100ms",
        "RestartUSecNext": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RootEphemeral": "no",
        "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "RuntimeRandomizedExtraUSec": "0",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "SetLoginEnvironment": "no",
        "Slice": "system.slice",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardOutput": "journal",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StartupMemoryHigh": "infinity",
        "StartupMemoryLow": "0",
        "StartupMemoryMax": "infinity",
        "StartupMemorySwapMax": "infinity",
        "StartupMemoryZSwapMax": "infinity",
        "StateChangeTimestamp": "Thu 2024-10-10 08:24:25 EDT",
        "StateChangeTimestampMonotonic": "1207512227",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "running",
        "SuccessAction": "none",
        "SurviveFinalKillSignal": "no",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "2147483646",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "32",
        "TasksMax": "22362",
        "TimeoutAbortUSec": "1min 30s",
        "TimeoutCleanUSec": "infinity",
        "TimeoutStartFailureMode": "terminate",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopFailureMode": "terminate",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "Type": "notify",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFilePreset": "disabled",
        "UnitFileState": "enabled",
        "UtmpMode": "init",
        "WantedBy": "multi-user.target",
        "WatchdogSignal": "6",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.ha_cluster : Regenerate pcsd TLS certificate and key] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:11
Thursday 10 October 2024  08:25:46 -0400 (0:00:00.795)       0:00:07.939 ****** 
skipping: [managed-node3] => (item=/var/lib/pcsd/pcsd.key)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "false_condition": "ha_cluster_regenerate_keys",
    "item": "/var/lib/pcsd/pcsd.key",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=/var/lib/pcsd/pcsd.crt)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "false_condition": "ha_cluster_regenerate_keys",
    "item": "/var/lib/pcsd/pcsd.crt",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => {
    "changed": false
}

MSG:

All items skipped

TASK [fedora.linux_system_roles.ha_cluster : Get the stat of /var/lib/pcsd] ****
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:25
Thursday 10 October 2024  08:25:46 -0400 (0:00:00.018)       0:00:07.957 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_pcsd_certificates | length > 0",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Allow certmonger to write into pcsd's certificate directory] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:30
Thursday 10 October 2024  08:25:46 -0400 (0:00:00.033)       0:00:07.990 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_pcsd_certificates | length > 0",
    "skip_reason": "Conditional result was False"
}

TASK [Ensure the name of ha_cluster_pcsd_certificates is /var/lib/pcsd/pcsd; Create certificates using the certificate role] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:37
Thursday 10 October 2024  08:25:46 -0400 (0:00:00.032)       0:00:08.023 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_pcsd_certificates | length > 0",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Set pcsd's certificate directory back to cluster_var_lib_t] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:49
Thursday 10 October 2024  08:25:46 -0400 (0:00:00.033)       0:00:08.056 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_pcsd_certificates | length > 0",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Distribute pcsd TLS private key] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:64
Thursday 10 October 2024  08:25:46 -0400 (0:00:00.036)       0:00:08.092 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_pcsd_public_key_src is string",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Distribute pcsd TLS certificate] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:71
Thursday 10 October 2024  08:25:46 -0400 (0:00:00.053)       0:00:08.146 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_pcsd_public_key_src is string",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Distribute pcs_settings.conf] *****
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:79
Thursday 10 October 2024  08:25:46 -0400 (0:00:00.058)       0:00:08.204 ****** 
changed: [managed-node3] => {
    "changed": true,
    "checksum": "b504e1b9c9aa23803dd6f95e66c757088b08551d",
    "dest": "/var/lib/pcsd/pcs_settings.conf",
    "gid": 0,
    "group": "root",
    "md5sum": "087ff556d850518c8fff5ad1179d8817",
    "mode": "0644",
    "owner": "root",
    "secontext": "system_u:object_r:cluster_var_lib_t:s0",
    "size": 359,
    "src": "/root/.ansible/tmp/ansible-tmp-1728563146.5317867-46183-145179813102995/.source.conf",
    "state": "file",
    "uid": 0
}

TASK [fedora.linux_system_roles.ha_cluster : Start pcsd with updated config files and configure it to start on boot] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:88
Thursday 10 October 2024  08:25:47 -0400 (0:00:00.727)       0:00:08.932 ****** 
changed: [managed-node3] => {
    "changed": true,
    "enabled": true,
    "name": "pcsd",
    "state": "started",
    "status": {
        "AccessSELinuxContext": "system_u:object_r:cluster_unit_file_t:s0",
        "ActiveEnterTimestamp": "Thu 2024-10-10 08:24:25 EDT",
        "ActiveEnterTimestampMonotonic": "1207512227",
        "ActiveExitTimestamp": "Thu 2024-10-10 08:25:46 EDT",
        "ActiveExitTimestampMonotonic": "1288292518",
        "ActiveState": "inactive",
        "After": "systemd-journald.socket network-online.target sysinit.target pcsd-ruby.service basic.target system.slice",
        "AllowIsolate": "no",
        "AssertResult": "yes",
        "AssertTimestamp": "Thu 2024-10-10 08:24:24 EDT",
        "AssertTimestampMonotonic": "1206731333",
        "Before": "multi-user.target shutdown.target",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "yes",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "yes",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore",
        "CleanResult": "success",
        "CollectMode": "inactive",
        "ConditionResult": "yes",
        "ConditionTimestamp": "Thu 2024-10-10 08:24:24 EDT",
        "ConditionTimestampMonotonic": "1206731330",
        "ConfigurationDirectoryMode": "0755",
        "Conflicts": "shutdown.target",
        "ConsistsOf": "pcsd-ruby.service",
        "ControlGroupId": "0",
        "ControlPID": "0",
        "CoredumpFilter": "0x33",
        "CoredumpReceive": "no",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "DefaultStartupMemoryLow": "0",
        "Delegate": "no",
        "Description": "PCS GUI and remote configuration interface",
        "DevicePolicy": "auto",
        "Documentation": "\"man:pcsd(8)\" \"man:pcs(8)\"",
        "DynamicUser": "no",
        "EffectiveMemoryHigh": "3702874112",
        "EffectiveMemoryMax": "3702874112",
        "EffectiveTasksMax": "22362",
        "EnvironmentFiles": "/etc/sysconfig/pcsd (ignore_errors=no)",
        "ExecMainCode": "1",
        "ExecMainExitTimestamp": "Thu 2024-10-10 08:25:46 EDT",
        "ExecMainExitTimestampMonotonic": "1288394339",
        "ExecMainHandoffTimestamp": "Thu 2024-10-10 08:24:24 EDT",
        "ExecMainHandoffTimestampMonotonic": "1206745416",
        "ExecMainPID": "114677",
        "ExecMainStartTimestamp": "Thu 2024-10-10 08:24:24 EDT",
        "ExecMainStartTimestampMonotonic": "1206732638",
        "ExecMainStatus": "0",
        "ExecStart": "{ path=/usr/sbin/pcsd ; argv[]=/usr/sbin/pcsd ; ignore_errors=no ; start_time=[Thu 2024-10-10 08:24:24 EDT] ; stop_time=[Thu 2024-10-10 08:25:46 EDT] ; pid=114677 ; code=exited ; status=0 }",
        "ExecStartEx": "{ path=/usr/sbin/pcsd ; argv[]=/usr/sbin/pcsd ; flags= ; start_time=[Thu 2024-10-10 08:24:24 EDT] ; stop_time=[Thu 2024-10-10 08:25:46 EDT] ; pid=114677 ; code=exited ; status=0 }",
        "ExitType": "main",
        "ExtensionImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FileDescriptorStorePreserve": "restart",
        "FinalKillSignal": "9",
        "FragmentPath": "/usr/lib/systemd/system/pcsd.service",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOReadBytes": "[not set]",
        "IOReadOperations": "[not set]",
        "IOSchedulingClass": "2",
        "IOSchedulingPriority": "4",
        "IOWeight": "[not set]",
        "IOWriteBytes": "[not set]",
        "IOWriteOperations": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "[no data]",
        "IPEgressPackets": "[no data]",
        "IPIngressBytes": "[no data]",
        "IPIngressPackets": "[no data]",
        "Id": "pcsd.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestamp": "Thu 2024-10-10 08:25:46 EDT",
        "InactiveEnterTimestampMonotonic": "1288397095",
        "InactiveExitTimestamp": "Thu 2024-10-10 08:24:24 EDT",
        "InactiveExitTimestampMonotonic": "1206734418",
        "InvocationID": "aa68975f714248cdbf40efb72dc43515",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "mixed",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "8388608",
        "LimitMEMLOCKSoft": "8388608",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "524288",
        "LimitNOFILESoft": "1024",
        "LimitNPROC": "13976",
        "LimitNPROCSoft": "13976",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "13976",
        "LimitSIGPENDINGSoft": "13976",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadState": "loaded",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "ManagedOOMMemoryPressure": "auto",
        "ManagedOOMMemoryPressureLimit": "0",
        "ManagedOOMPreference": "none",
        "ManagedOOMSwap": "auto",
        "MemoryAccounting": "yes",
        "MemoryAvailable": "3098513408",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryKSM": "no",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemoryPeak": "[not set]",
        "MemoryPressureThresholdUSec": "200ms",
        "MemoryPressureWatch": "auto",
        "MemorySwapCurrent": "[not set]",
        "MemorySwapMax": "infinity",
        "MemorySwapPeak": "[not set]",
        "MemoryZSwapCurrent": "[not set]",
        "MemoryZSwapMax": "infinity",
        "MemoryZSwapWriteback": "yes",
        "MountAPIVFS": "no",
        "MountImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAPolicy": "n/a",
        "Names": "pcsd.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "main",
        "OOMPolicy": "stop",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "OnSuccessJobMode": "fail",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateIPC": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProcSubset": "all",
        "ProtectClock": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectHostname": "no",
        "ProtectKernelLogs": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectProc": "default",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "ReloadResult": "success",
        "ReloadSignal": "1",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Requires": "network-online.target sysinit.target pcsd-ruby.service system.slice",
        "Restart": "no",
        "RestartKillSignal": "15",
        "RestartMaxDelayUSec": "infinity",
        "RestartMode": "normal",
        "RestartSteps": "0",
        "RestartUSec": "100ms",
        "RestartUSecNext": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RootEphemeral": "no",
        "RootImagePolicy": "root=verity+signed+encrypted+unprotected+absent:usr=verity+signed+encrypted+unprotected+absent:home=encrypted+unprotected+absent:srv=encrypted+unprotected+absent:tmp=encrypted+unprotected+absent:var=encrypted+unprotected+absent",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "RuntimeRandomizedExtraUSec": "0",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "SetLoginEnvironment": "no",
        "Slice": "system.slice",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardOutput": "journal",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StartupMemoryHigh": "infinity",
        "StartupMemoryLow": "0",
        "StartupMemoryMax": "infinity",
        "StartupMemorySwapMax": "infinity",
        "StartupMemoryZSwapMax": "infinity",
        "StateChangeTimestamp": "Thu 2024-10-10 08:25:46 EDT",
        "StateChangeTimestampMonotonic": "1288397095",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SurviveFinalKillSignal": "no",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "2147483646",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22362",
        "TimeoutAbortUSec": "1min 30s",
        "TimeoutCleanUSec": "infinity",
        "TimeoutStartFailureMode": "terminate",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopFailureMode": "terminate",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "Type": "notify",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFilePreset": "disabled",
        "UnitFileState": "enabled",
        "UtmpMode": "init",
        "WantedBy": "multi-user.target",
        "WatchdogSignal": "6",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.ha_cluster : Configure firewall] ***************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:49
Thursday 10 October 2024  08:25:49 -0400 (0:00:01.911)       0:00:10.843 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_cluster_present | bool or ha_cluster_qnetd.present | d(false)",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Configure selinux] ****************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:52
Thursday 10 October 2024  08:25:49 -0400 (0:00:00.034)       0:00:10.877 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_cluster_present | bool or ha_cluster_qnetd.present | d(false)",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Install cluster packages] *********
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:58
Thursday 10 October 2024  08:25:49 -0400 (0:00:00.034)       0:00:10.912 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_cluster_present | bool",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Distribute fence-virt authkey] ****
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:64
Thursday 10 October 2024  08:25:49 -0400 (0:00:00.030)       0:00:10.943 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_cluster_present | bool",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Configure SBD] ********************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:69
Thursday 10 October 2024  08:25:49 -0400 (0:00:00.032)       0:00:10.975 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_cluster_present | bool",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Configure corosync] ***************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:72
Thursday 10 October 2024  08:25:49 -0400 (0:00:00.029)       0:00:11.005 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_cluster_present | bool",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Cluster auth] *********************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:75
Thursday 10 October 2024  08:25:49 -0400 (0:00:00.030)       0:00:11.036 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_cluster_present | bool",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Distribute cluster shared keys] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:80
Thursday 10 October 2024  08:25:49 -0400 (0:00:00.030)       0:00:11.066 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_cluster_present | bool",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Enable or disable cluster services on boot] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:86
Thursday 10 October 2024  08:25:49 -0400 (0:00:00.029)       0:00:11.096 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_cluster_present | bool",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Start the cluster and reload corosync.conf] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:89
Thursday 10 October 2024  08:25:49 -0400 (0:00:00.029)       0:00:11.126 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_cluster_present | bool",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Create and push CIB] **************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:92
Thursday 10 October 2024  08:25:49 -0400 (0:00:00.029)       0:00:11.155 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_cluster_present | bool",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Remove cluster configuration] *****
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:101
Thursday 10 October 2024  08:25:49 -0400 (0:00:00.031)       0:00:11.186 ****** 
included: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-destroy-pcs-0.10.yml for managed-node3

TASK [fedora.linux_system_roles.ha_cluster : Remove cluster configuration] *****
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-destroy-pcs-0.10.yml:9
Thursday 10 October 2024  08:25:49 -0400 (0:00:00.019)       0:00:11.205 ****** 
changed: [managed-node3] => (item=/etc/corosync/corosync.conf) => {
    "ansible_loop_var": "item",
    "changed": true,
    "cmd": [
        "pcs",
        "cluster",
        "destroy"
    ],
    "delta": "0:00:03.178188",
    "end": "2024-10-10 08:25:53.209624",
    "item": "/etc/corosync/corosync.conf",
    "rc": 0,
    "start": "2024-10-10 08:25:50.031436"
}

STDERR:

Shutting down pacemaker/corosync services...
Killing any remaining services...
Removing all cluster configuration files...
ok: [managed-node3] => (item=/var/lib/pacemaker/cib/cib.xml) => {
    "ansible_loop_var": "item",
    "changed": false,
    "cmd": [
        "pcs",
        "cluster",
        "destroy"
    ],
    "delta": null,
    "end": null,
    "item": "/var/lib/pacemaker/cib/cib.xml",
    "rc": 0,
    "start": null
}

STDOUT:

skipped, since /var/lib/pacemaker/cib/cib.xml does not exist


MSG:

Did not run command since '/var/lib/pacemaker/cib/cib.xml' does not exist

TASK [fedora.linux_system_roles.ha_cluster : Remove fence-virt authkey] ********
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:104
Thursday 10 October 2024  08:25:53 -0400 (0:00:04.422)       0:00:15.628 ****** 
changed: [managed-node3] => {
    "changed": true,
    "path": "/etc/cluster/fence_xvm.key",
    "state": "absent"
}

TASK [fedora.linux_system_roles.ha_cluster : Configure qnetd] ******************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:109
Thursday 10 October 2024  08:25:54 -0400 (0:00:00.720)       0:00:16.348 ****** 
included: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml for managed-node3

TASK [fedora.linux_system_roles.ha_cluster : Remove qnetd configuration] *******
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml:3
Thursday 10 October 2024  08:25:54 -0400 (0:00:00.020)       0:00:16.369 ****** 
changed: [managed-node3] => {
    "changed": true,
    "cmd": [
        "pcs",
        "--force",
        "--",
        "qdevice",
        "destroy",
        "net"
    ],
    "delta": "0:00:01.423375",
    "end": "2024-10-10 08:25:56.605312",
    "rc": 0,
    "start": "2024-10-10 08:25:55.181937"
}

STDERR:

Stopping quorum device...
quorum device stopped
quorum device disabled
Quorum device 'net' configuration files removed

TASK [fedora.linux_system_roles.ha_cluster : Setup qnetd] **********************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml:16
Thursday 10 October 2024  08:25:56 -0400 (0:00:02.020)       0:00:18.389 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_qnetd.present | d(false)",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Enable or disable qnetd service on boot] ***
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml:26
Thursday 10 October 2024  08:25:56 -0400 (0:00:00.039)       0:00:18.428 ****** 
skipping: [managed-node3] => {
    "changed": false,
    "false_condition": "ha_cluster_qnetd.present | d(false)",
    "skip_reason": "Conditional result was False"
}

TASK [Stat corosync.conf] ******************************************************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cluster_destroy.yml:28
Thursday 10 October 2024  08:25:56 -0400 (0:00:00.036)       0:00:18.465 ****** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "exists": false
    }
}

TASK [Stat cib.xml] ************************************************************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cluster_destroy.yml:33
Thursday 10 October 2024  08:25:57 -0400 (0:00:00.328)       0:00:18.794 ****** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "exists": false
    }
}

TASK [Stat fence_xvm.key] ******************************************************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cluster_destroy.yml:38
Thursday 10 October 2024  08:25:57 -0400 (0:00:00.382)       0:00:19.176 ****** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "exists": false
    }
}

TASK [Check the files do not exist] ********************************************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cluster_destroy.yml:43
Thursday 10 October 2024  08:25:57 -0400 (0:00:00.320)       0:00:19.497 ****** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Export cluster configuration] ********************************************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cluster_destroy.yml:50
Thursday 10 October 2024  08:25:57 -0400 (0:00:00.019)       0:00:19.516 ****** 
ok: [managed-node3] => {
    "changed": false,
    "ha_cluster": {
        "ha_cluster_cluster_present": false
    }
}

TASK [Check exported configuration] ********************************************
task path: /tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cluster_destroy.yml:54
Thursday 10 October 2024  08:25:58 -0400 (0:00:00.406)       0:00:19.922 ****** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

PLAY RECAP *********************************************************************
managed-node3              : ok=42   changed=6    unreachable=0    failed=0    skipped=36   rescued=0    ignored=0   

Thursday 10 October 2024  08:25:58 -0400 (0:00:00.024)       0:00:19.947 ****** 
=============================================================================== 
fedora.linux_system_roles.ha_cluster : Remove cluster configuration ----- 4.42s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/cluster-destroy-pcs-0.10.yml:9 
fedora.linux_system_roles.ha_cluster : Remove qnetd configuration ------- 2.02s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/pcs-qnetd.yml:3 
fedora.linux_system_roles.ha_cluster : Start pcsd with updated config files and configure it to start on boot --- 1.91s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:88 
Gathering Facts --------------------------------------------------------- 1.19s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cluster_destroy.yml:9 
fedora.linux_system_roles.ha_cluster : Fetch pcsd capabilities ---------- 0.99s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:155 
fedora.linux_system_roles.ha_cluster : Fetch pcs capabilities ----------- 0.92s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:141 
fedora.linux_system_roles.ha_cluster : Stop pcsd ------------------------ 0.80s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:6 
fedora.linux_system_roles.ha_cluster : Install role essential packages --- 0.79s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:11 
fedora.linux_system_roles.ha_cluster : Distribute pcs_settings.conf ----- 0.73s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/configure-shell.yml:79 
fedora.linux_system_roles.ha_cluster : Remove fence-virt authkey -------- 0.72s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:104 
fedora.linux_system_roles.ha_cluster : List active CentOS repositories --- 0.66s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml:3 
fedora.linux_system_roles.ha_cluster : Set hacluster password ----------- 0.58s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:38 
fedora.linux_system_roles.ha_cluster : Generate a password hash --------- 0.45s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:26 
fedora.linux_system_roles.ha_cluster : Check if system is ostree -------- 0.41s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:22 
Export cluster configuration -------------------------------------------- 0.41s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cluster_destroy.yml:50 
Stat cib.xml ------------------------------------------------------------ 0.38s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cluster_destroy.yml:33 
Stat corosync.conf ------------------------------------------------------ 0.33s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cluster_destroy.yml:28 
Stat fence_xvm.key ------------------------------------------------------ 0.32s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_cluster_destroy.yml:38 
fedora.linux_system_roles.ha_cluster : Fail if passwords are not specified --- 0.08s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:43 
fedora.linux_system_roles.ha_cluster : Parse pcsd capabilities ---------- 0.06s
/tmp/collections-kFV/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:163