[DEPRECATION WARNING]: ANSIBLE_COLLECTIONS_PATHS option, does not fit var 
naming standard, use the singular form ANSIBLE_COLLECTIONS_PATH instead. This 
feature will be removed from ansible-core in version 2.19. Deprecation warnings
 can be disabled by setting deprecation_warnings=False in ansible.cfg.
ansible-playbook [core 2.17.2]
  config file = None
  configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
  ansible python module location = /usr/local/lib/python3.12/site-packages/ansible
  ansible collection location = /tmp/tmp.JVrTHQ6u36
  executable location = /usr/local/bin/ansible-playbook
  python version = 3.12.4 (main, Jul 17 2024, 00:00:00) [GCC 14.1.1 20240607 (Red Hat 14.1.1-5)] (/usr/bin/python3.12)
  jinja version = 3.1.4
  libyaml = True
No config file found; using defaults
running playbook inside collection fedora.linux_system_roles
redirecting (type: callback) ansible.builtin.debug to ansible.posix.debug
redirecting (type: callback) ansible.builtin.debug to ansible.posix.debug
redirecting (type: callback) ansible.builtin.profile_tasks to ansible.posix.profile_tasks
Skipping callback 'default', as we already have a stdout callback.
Skipping callback 'minimal', as we already have a stdout callback.
Skipping callback 'oneline', as we already have a stdout callback.

PLAYBOOK: tests_qnetd_disabled.yml *********************************************
2 plays in /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_qnetd_disabled.yml

PLAY [all] *********************************************************************

TASK [Include vault variables] *************************************************
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_qnetd_disabled.yml:5
Monday 29 July 2024  07:42:12 -0400 (0:00:00.007)       0:00:00.007 *********** 
ok: [managed_node1] => {
    "ansible_facts": {
        "ha_cluster_hacluster_password": {
            "__ansible_vault": "$ANSIBLE_VAULT;1.1;AES256\n31303833633366333561656439323930303361333161363239346166656537323933313436\n3432386236656563343237306335323637396239616230353561330a313731623238393238\n62343064666336643930663239383936616465643134646536656532323461356237646133\n3761616633323839633232353637366266350a313163633236376666653238633435306565\n3264623032333736393535663833\n"
        }
    },
    "ansible_included_var_files": [
        "/var/ARTIFACTS/work-generall14g4j2x/plans/general/tree/tests/vars/vault-variables.yml"
    ],
    "changed": false
}

PLAY [Test qnetd setup with qnetd not starting on boot] ************************

TASK [Gathering Facts] *********************************************************
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_qnetd_disabled.yml:9
Monday 29 July 2024  07:42:12 -0400 (0:00:00.033)       0:00:00.041 *********** 
[WARNING]: Platform linux on host managed_node1 is using the discovered Python
interpreter at /usr/bin/python3.12, but future installation of another Python
interpreter could change the meaning of that path. See
https://docs.ansible.com/ansible-
core/2.17/reference_appendices/interpreter_discovery.html for more information.
ok: [managed_node1]

TASK [Set up test environment] *************************************************
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_qnetd_disabled.yml:21
Monday 29 July 2024  07:42:13 -0400 (0:00:01.079)       0:00:01.120 *********** 
included: fedora.linux_system_roles.ha_cluster for managed_node1

TASK [fedora.linux_system_roles.ha_cluster : Set node name to 'localhost' for single-node clusters] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:9
Monday 29 July 2024  07:42:13 -0400 (0:00:00.026)       0:00:01.147 *********** 
ok: [managed_node1] => {
    "ansible_facts": {
        "inventory_hostname": "localhost"
    },
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Ensure facts used by tests] *******
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:14
Monday 29 July 2024  07:42:13 -0400 (0:00:00.019)       0:00:01.166 *********** 
skipping: [managed_node1] => {
    "changed": false,
    "false_condition": "'distribution' not in ansible_facts",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Check if system is ostree] ********
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:22
Monday 29 July 2024  07:42:13 -0400 (0:00:00.013)       0:00:01.180 *********** 
ok: [managed_node1] => {
    "changed": false,
    "stat": {
        "exists": false
    }
}

TASK [fedora.linux_system_roles.ha_cluster : Set flag to indicate system is ostree] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:27
Monday 29 July 2024  07:42:14 -0400 (0:00:00.403)       0:00:01.583 *********** 
ok: [managed_node1] => {
    "ansible_facts": {
        "__ha_cluster_is_ostree": false
    },
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Do not try to enable RHEL repositories] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:32
Monday 29 July 2024  07:42:14 -0400 (0:00:00.023)       0:00:01.606 *********** 
skipping: [managed_node1] => {
    "changed": false,
    "false_condition": "ansible_distribution == 'RedHat'",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Copy nss-altfiles ha_cluster users to /etc/passwd] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:41
Monday 29 July 2024  07:42:14 -0400 (0:00:00.012)       0:00:01.619 *********** 
skipping: [managed_node1] => {
    "changed": false,
    "false_condition": "__ha_cluster_is_ostree | d(false)",
    "skip_reason": "Conditional result was False"
}

TASK [Clean up test environment for qnetd / qdevice] ***************************
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_qnetd_disabled.yml:26
Monday 29 July 2024  07:42:14 -0400 (0:00:00.019)       0:00:01.638 *********** 
included: fedora.linux_system_roles.ha_cluster for managed_node1

TASK [fedora.linux_system_roles.ha_cluster : Make sure qnetd is not installed] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_cleanup_qnetd.yml:9
Monday 29 July 2024  07:42:14 -0400 (0:00:00.024)       0:00:01.662 *********** 
changed: [managed_node1] => {
    "changed": true,
    "rc": 0,
    "results": [
        "Removed: corosync-qnetd-3.0.3-6.el10.x86_64"
    ]
}

TASK [fedora.linux_system_roles.ha_cluster : Make sure qnetd config files are not present] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_cleanup_qnetd.yml:16
Monday 29 July 2024  07:42:15 -0400 (0:00:01.399)       0:00:03.062 *********** 
ok: [managed_node1] => {
    "changed": false,
    "path": "/etc/corosync/qnetd",
    "state": "absent"
}

TASK [Run HA Cluster role] *****************************************************
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_qnetd_disabled.yml:31
Monday 29 July 2024  07:42:15 -0400 (0:00:00.430)       0:00:03.493 *********** 
included: fedora.linux_system_roles.ha_cluster for managed_node1

TASK [fedora.linux_system_roles.ha_cluster : Set platform/version specific variables] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:3
Monday 29 July 2024  07:42:16 -0400 (0:00:00.044)       0:00:03.537 *********** 
included: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml for managed_node1

TASK [fedora.linux_system_roles.ha_cluster : Ensure ansible_facts used by role] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:2
Monday 29 July 2024  07:42:16 -0400 (0:00:00.019)       0:00:03.557 *********** 
skipping: [managed_node1] => {
    "changed": false,
    "false_condition": "__ha_cluster_required_facts | difference(ansible_facts.keys() | list) | length > 0",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Check if system is ostree] ********
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:10
Monday 29 July 2024  07:42:16 -0400 (0:00:00.021)       0:00:03.578 *********** 
skipping: [managed_node1] => {
    "changed": false,
    "false_condition": "not __ha_cluster_is_ostree is defined",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Set flag to indicate system is ostree] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:15
Monday 29 July 2024  07:42:16 -0400 (0:00:00.015)       0:00:03.594 *********** 
skipping: [managed_node1] => {
    "changed": false,
    "false_condition": "not __ha_cluster_is_ostree is defined",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Set platform/version specific variables] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:19
Monday 29 July 2024  07:42:16 -0400 (0:00:00.015)       0:00:03.609 *********** 
ok: [managed_node1] => (item=RedHat.yml) => {
    "ansible_facts": {
        "__ha_cluster_cloud_agents_packages": [],
        "__ha_cluster_fence_agent_packages_default": "{{ ['fence-agents-all'] + (['fence-virt'] if ansible_architecture == 'x86_64' else []) }}",
        "__ha_cluster_fullstack_node_packages": [
            "corosync",
            "libknet1-plugins-all",
            "resource-agents",
            "pacemaker",
            "openssl"
        ],
        "__ha_cluster_pcs_provider": "pcs-0.10",
        "__ha_cluster_qdevice_node_packages": [
            "corosync-qdevice",
            "bash",
            "coreutils",
            "curl",
            "grep",
            "nss-tools",
            "openssl",
            "sed"
        ],
        "__ha_cluster_repos": [],
        "__ha_cluster_role_essential_packages": [
            "pcs",
            "corosync-qnetd"
        ],
        "__ha_cluster_sbd_packages": [
            "sbd"
        ],
        "__ha_cluster_services": [
            "corosync",
            "corosync-qdevice",
            "pacemaker"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/RedHat.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml"
}
skipping: [managed_node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "false_condition": "__vars_file is file",
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed_node1] => (item=CentOS_10.yml) => {
    "ansible_facts": {
        "__ha_cluster_cloud_agents_packages": [
            "resource-agents-cloud",
            "fence-agents-aliyun",
            "fence-agents-aws",
            "fence-agents-azure-arm",
            "fence-agents-compute",
            "fence-agents-gce",
            "fence-agents-ibm-powervs",
            "fence-agents-ibm-vpc",
            "fence-agents-kubevirt",
            "fence-agents-openstack"
        ],
        "__ha_cluster_repos": [
            {
                "id": "highavailability",
                "name": "HighAvailability"
            },
            {
                "id": "resilientstorage",
                "name": "ResilientStorage"
            }
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/CentOS_10.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_10.yml"
}
ok: [managed_node1] => (item=CentOS_10.yml) => {
    "ansible_facts": {
        "__ha_cluster_cloud_agents_packages": [
            "resource-agents-cloud",
            "fence-agents-aliyun",
            "fence-agents-aws",
            "fence-agents-azure-arm",
            "fence-agents-compute",
            "fence-agents-gce",
            "fence-agents-ibm-powervs",
            "fence-agents-ibm-vpc",
            "fence-agents-kubevirt",
            "fence-agents-openstack"
        ],
        "__ha_cluster_repos": [
            {
                "id": "highavailability",
                "name": "HighAvailability"
            },
            {
                "id": "resilientstorage",
                "name": "ResilientStorage"
            }
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/CentOS_10.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_10.yml"
}

TASK [fedora.linux_system_roles.ha_cluster : Set Linux Pacemaker shell specific variables] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:34
Monday 29 July 2024  07:42:16 -0400 (0:00:00.039)       0:00:03.649 *********** 
ok: [managed_node1] => {
    "ansible_facts": {},
    "ansible_included_var_files": [
        "/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/vars/shell_pcs.yml"
    ],
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Enable package repositories] ******
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:6
Monday 29 July 2024  07:42:16 -0400 (0:00:00.015)       0:00:03.664 *********** 
included: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-package-repositories.yml for managed_node1

TASK [fedora.linux_system_roles.ha_cluster : Find platform/version specific tasks to enable repositories] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-package-repositories.yml:3
Monday 29 July 2024  07:42:16 -0400 (0:00:00.020)       0:00:03.684 *********** 
ok: [managed_node1] => (item=RedHat.yml) => {
    "ansible_facts": {
        "__ha_cluster_enable_repo_tasks_file": "/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/RedHat.yml"
    },
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml"
}
ok: [managed_node1] => (item=CentOS.yml) => {
    "ansible_facts": {
        "__ha_cluster_enable_repo_tasks_file": "/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml"
    },
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml"
}
skipping: [managed_node1] => (item=CentOS_10.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "false_condition": "__ha_cluster_enable_repo_tasks_file_candidate is file",
    "item": "CentOS_10.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed_node1] => (item=CentOS_10.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "false_condition": "__ha_cluster_enable_repo_tasks_file_candidate is file",
    "item": "CentOS_10.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Run platform/version specific tasks to enable repositories] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-package-repositories.yml:21
Monday 29 July 2024  07:42:16 -0400 (0:00:00.039)       0:00:03.724 *********** 
included: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml for managed_node1

TASK [fedora.linux_system_roles.ha_cluster : List active CentOS repositories] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml:3
Monday 29 July 2024  07:42:16 -0400 (0:00:00.030)       0:00:03.755 *********** 
ok: [managed_node1] => {
    "changed": false,
    "cmd": [
        "dnf",
        "repolist"
    ],
    "delta": "0:00:00.192431",
    "end": "2024-07-29 07:42:16.787617",
    "rc": 0,
    "start": "2024-07-29 07:42:16.595186"
}

STDOUT:

repo id                 repo name
appstream               CentOS Stream 10 - AppStream
baseos                  CentOS Stream 10 - BaseOS
beaker-client           Beaker Client - RedHatEnterpriseLinux9
beaker-harness          Beaker harness
beakerlib-libraries     Copr repo for beakerlib-libraries owned by bgoncalv
highavailability        CentOS Stream 10 - HighAvailability

TASK [fedora.linux_system_roles.ha_cluster : Enable CentOS repositories] *******
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml:10
Monday 29 July 2024  07:42:16 -0400 (0:00:00.594)       0:00:04.349 *********** 
skipping: [managed_node1] => (item={'id': 'highavailability', 'name': 'HighAvailability'})  => {
    "ansible_loop_var": "item",
    "changed": false,
    "false_condition": "item.id not in __ha_cluster_repolist.stdout",
    "item": {
        "id": "highavailability",
        "name": "HighAvailability"
    },
    "skip_reason": "Conditional result was False"
}
skipping: [managed_node1] => (item={'id': 'resilientstorage', 'name': 'ResilientStorage'})  => {
    "ansible_loop_var": "item",
    "changed": false,
    "false_condition": "item.name != \"ResilientStorage\" or ha_cluster_enable_repos_resilient_storage",
    "item": {
        "id": "resilientstorage",
        "name": "ResilientStorage"
    },
    "skip_reason": "Conditional result was False"
}
skipping: [managed_node1] => {
    "changed": false
}

MSG:

All items skipped

TASK [fedora.linux_system_roles.ha_cluster : Install role essential packages] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:11
Monday 29 July 2024  07:42:16 -0400 (0:00:00.018)       0:00:04.368 *********** 
changed: [managed_node1] => {
    "changed": true,
    "rc": 0,
    "results": [
        "Installed: corosync-qnetd-3.0.3-6.el10.x86_64"
    ]
}

TASK [fedora.linux_system_roles.ha_cluster : Check and prepare role variables] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:17
Monday 29 July 2024  07:42:18 -0400 (0:00:01.496)       0:00:05.864 *********** 
included: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml for managed_node1

TASK [fedora.linux_system_roles.ha_cluster : Discover cluster node names] ******
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:3
Monday 29 July 2024  07:42:18 -0400 (0:00:00.036)       0:00:05.901 *********** 
ok: [managed_node1] => {
    "ansible_facts": {
        "__ha_cluster_node_name": "localhost"
    },
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Collect cluster node names] *******
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:7
Monday 29 July 2024  07:42:18 -0400 (0:00:00.022)       0:00:05.923 *********** 
ok: [managed_node1] => {
    "ansible_facts": {
        "__ha_cluster_all_node_names": [
            "localhost"
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Fail if ha_cluster_node_options contains unknown or duplicate nodes] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:16
Monday 29 July 2024  07:42:18 -0400 (0:00:00.023)       0:00:05.946 *********** 
skipping: [managed_node1] => {
    "changed": false,
    "false_condition": "(\n  __nodes_from_options != (__nodes_from_options | unique)\n) or (\n  __nodes_from_options | difference(__ha_cluster_all_node_names)\n)\n",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Extract node options] *************
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:30
Monday 29 July 2024  07:42:18 -0400 (0:00:00.018)       0:00:05.965 *********** 
ok: [managed_node1] => {
    "ansible_facts": {
        "__ha_cluster_local_node": {}
    },
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Fail if passwords are not specified] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:43
Monday 29 July 2024  07:42:18 -0400 (0:00:00.024)       0:00:05.990 *********** 
skipping: [managed_node1] => (item=ha_cluster_hacluster_password)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "false_condition": "lookup(\"vars\", item, default=\"\") | string | length < 1",
    "item": "ha_cluster_hacluster_password",
    "skip_reason": "Conditional result was False"
}
skipping: [managed_node1] => {
    "changed": false
}

MSG:

All items skipped

TASK [fedora.linux_system_roles.ha_cluster : Fail if nodes do not have the same number of SBD devices specified] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:53
Monday 29 July 2024  07:42:18 -0400 (0:00:00.042)       0:00:06.033 *********** 
skipping: [managed_node1] => {
    "changed": false,
    "false_condition": "ha_cluster_cluster_present",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Fail if configuring qnetd on a cluster node] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:69
Monday 29 July 2024  07:42:18 -0400 (0:00:00.012)       0:00:06.045 *********** 
skipping: [managed_node1] => {
    "changed": false,
    "false_condition": "ha_cluster_cluster_present",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Fail if no valid level is specified for a fencing level] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:79
Monday 29 July 2024  07:42:18 -0400 (0:00:00.011)       0:00:06.057 *********** 
skipping: [managed_node1] => {
    "changed": false,
    "skipped_reason": "No items in the list"
}

TASK [fedora.linux_system_roles.ha_cluster : Fail if no target is specified for a fencing level] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:87
Monday 29 July 2024  07:42:18 -0400 (0:00:00.011)       0:00:06.069 *********** 
skipping: [managed_node1] => {
    "changed": false,
    "skipped_reason": "No items in the list"
}

TASK [fedora.linux_system_roles.ha_cluster : Extract qdevice settings] *********
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:101
Monday 29 July 2024  07:42:18 -0400 (0:00:00.011)       0:00:06.080 *********** 
ok: [managed_node1] => {
    "ansible_facts": {
        "__ha_cluster_qdevice_host": "",
        "__ha_cluster_qdevice_in_use": false,
        "__ha_cluster_qdevice_model": "",
        "__ha_cluster_qdevice_pcs_address": ""
    },
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Figure out if ATB needs to be enabled for SBD] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:110
Monday 29 July 2024  07:42:18 -0400 (0:00:00.029)       0:00:06.109 *********** 
ok: [managed_node1] => {
    "ansible_facts": {
        "__ha_cluster_sbd_needs_atb": false
    },
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Fail if SBD needs ATB enabled and the user configured ATB to be disabled] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:120
Monday 29 July 2024  07:42:18 -0400 (0:00:00.023)       0:00:06.132 *********** 
skipping: [managed_node1] => {
    "changed": false,
    "false_condition": "__ha_cluster_sbd_needs_atb | bool",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Fail if ha_cluster_pcsd_public_key_src and ha_cluster_pcsd_private_key_src are set along with ha_cluster_pcsd_certificates] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:127
Monday 29 July 2024  07:42:18 -0400 (0:00:00.018)       0:00:06.151 *********** 
skipping: [managed_node1] => {
    "changed": false,
    "false_condition": "ha_cluster_pcsd_public_key_src is not none",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Fetch pcs capabilities] ***********
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:141
Monday 29 July 2024  07:42:18 -0400 (0:00:00.015)       0:00:06.167 *********** 
ok: [managed_node1] => {
    "changed": false,
    "cmd": [
        "pcs",
        "--version",
        "--full"
    ],
    "delta": "0:00:01.610929",
    "end": "2024-07-29 07:42:20.538480",
    "rc": 0,
    "start": "2024-07-29 07:42:18.927551"
}

STDOUT:

0.12.0a1
booth booth.enable-authfile.unset cluster.config.backup-local cluster.config.restore-cluster cluster.config.restore-local cluster.config.uuid cluster.create cluster.create.enable cluster.create.local cluster.create.no-keys-sync cluster.create.separated-name-and-address cluster.create.start cluster.create.start.wait cluster.create.transport.knet cluster.create.transport.udp-udpu cluster.create.transport.udp-udpu.no-rrp cluster.destroy cluster.destroy.all cluster.report cluster.verify corosync.authkey.update corosync.config.get corosync.config.get.struct corosync.config.reload corosync.config.sync-to-local-cluster corosync.config.update corosync.link.add corosync.link.remove corosync.link.remove.list corosync.link.update corosync.qdevice corosync.qdevice.model.net corosync.quorum corosync.quorum.device corosync.quorum.device.client.model.net.certificates.local corosync.quorum.device.heuristics corosync.quorum.device.model.net corosync.quorum.device.model.net.options_tls_and_kaptb corosync.quorum.set-expected-votes-runtime corosync.quorum.status corosync.quorum.unblock corosync.totem.block_unlisted_ips corosync.uidgid node.add node.add.enable node.add.separated-name-and-address node.add.start node.add.start.wait node.attributes node.attributes.set-list-for-node node.confirm-off node.fence node.guest node.kill node.maintenance node.maintenance.all node.maintenance.list node.maintenance.wait node.remote node.remote.onfail-demote node.remove node.remove-from-caches node.remove.list node.standby node.standby.all node.standby.list node.standby.wait node.start-stop-enable-disable node.start-stop-enable-disable.all node.start-stop-enable-disable.list node.start-stop-enable-disable.start-wait node.utilization node.utilization.set-list-for-node pcmk.acl.enable-disable pcmk.acl.group pcmk.acl.role pcmk.acl.role.create-with-permissions pcmk.acl.role.delete-with-users-groups pcmk.acl.user pcmk.alert pcmk.cib.checkpoints pcmk.cib.checkpoints.diff pcmk.cib.edit pcmk.cib.get pcmk.cib.get.scope pcmk.cib.roles.promoted-unpromoted pcmk.cib.set pcmk.constraint.colocation.set pcmk.constraint.colocation.set.options pcmk.constraint.colocation.simple pcmk.constraint.colocation.simple.options pcmk.constraint.colocation.simple.options.score pcmk.constraint.config.output-formats pcmk.constraint.hide-expired pcmk.constraint.location.simple pcmk.constraint.location.simple.options pcmk.constraint.location.simple.options.score pcmk.constraint.location.simple.resource-regexp pcmk.constraint.location.simple.rule pcmk.constraint.location.simple.rule.node-attr-type-number pcmk.constraint.location.simple.rule.options pcmk.constraint.location.simple.rule.rule-add-remove pcmk.constraint.location.simple.rule.rule-as-one-argument pcmk.constraint.no-autocorrect pcmk.constraint.order.set pcmk.constraint.order.set.options pcmk.constraint.order.simple pcmk.constraint.order.simple.options pcmk.constraint.ticket.set pcmk.constraint.ticket.set.options pcmk.constraint.ticket.simple pcmk.constraint.ticket.simple.constraint-id pcmk.properties.cluster pcmk.properties.cluster.config.output-formats pcmk.properties.cluster.defaults pcmk.properties.cluster.describe pcmk.properties.cluster.describe.output-formats pcmk.properties.cluster.set_properties.iso8601duration pcmk.properties.operation-defaults pcmk.properties.operation-defaults.config.output-formats pcmk.properties.operation-defaults.multiple pcmk.properties.operation-defaults.rule pcmk.properties.operation-defaults.rule-rsc-op pcmk.properties.operation-defaults.rule.hide-expired pcmk.properties.operation-defaults.rule.node-attr-type-number pcmk.properties.resource-defaults pcmk.properties.resource-defaults.config.output-formats pcmk.properties.resource-defaults.multiple pcmk.properties.resource-defaults.rule pcmk.properties.resource-defaults.rule-rsc-op pcmk.properties.resource-defaults.rule.hide-expired pcmk.properties.resource-defaults.rule.node-attr-type-number pcmk.resource.ban-move-clear pcmk.resource.ban-move-clear.bundles pcmk.resource.ban-move-clear.clear-expired pcmk.resource.ban-move-clear.clone pcmk.resource.bundle pcmk.resource.bundle.container-docker pcmk.resource.bundle.container-docker.promoted-max pcmk.resource.bundle.container-podman pcmk.resource.bundle.container-podman.promoted-max pcmk.resource.bundle.container-rkt pcmk.resource.bundle.container-rkt.promoted-max pcmk.resource.bundle.reset pcmk.resource.bundle.wait pcmk.resource.cleanup pcmk.resource.cleanup.one-resource pcmk.resource.cleanup.strict pcmk.resource.clone pcmk.resource.clone.custom-id pcmk.resource.clone.meta-in-create pcmk.resource.clone.wait pcmk.resource.config.output-formats pcmk.resource.create pcmk.resource.create.clone.custom-id pcmk.resource.create.group.future pcmk.resource.create.in-existing-bundle pcmk.resource.create.meta pcmk.resource.create.meta.future pcmk.resource.create.no-master pcmk.resource.create.operations pcmk.resource.create.operations.onfail-demote pcmk.resource.create.promotable pcmk.resource.create.promotable.custom-id pcmk.resource.create.wait pcmk.resource.debug pcmk.resource.delete pcmk.resource.disable.safe pcmk.resource.disable.safe.brief pcmk.resource.disable.safe.tag pcmk.resource.disable.simulate pcmk.resource.disable.simulate.brief pcmk.resource.disable.simulate.tag pcmk.resource.enable-disable pcmk.resource.enable-disable.list pcmk.resource.enable-disable.tag pcmk.resource.enable-disable.wait pcmk.resource.failcount pcmk.resource.group pcmk.resource.group.add-remove-list pcmk.resource.group.wait pcmk.resource.manage-unmanage pcmk.resource.manage-unmanage.list pcmk.resource.manage-unmanage.tag pcmk.resource.manage-unmanage.with-monitor pcmk.resource.move.autoclean pcmk.resource.move.autoclean.default pcmk.resource.promotable pcmk.resource.promotable.custom-id pcmk.resource.promotable.meta-in-create pcmk.resource.promotable.wait pcmk.resource.refresh pcmk.resource.refresh.one-resource pcmk.resource.refresh.strict pcmk.resource.relations pcmk.resource.relocate pcmk.resource.restart pcmk.resource.update pcmk.resource.update-meta pcmk.resource.update-meta.list pcmk.resource.update-meta.wait pcmk.resource.update-operations pcmk.resource.update-operations.onfail-demote pcmk.resource.update.meta pcmk.resource.update.operations pcmk.resource.update.operations.onfail-demote pcmk.resource.update.wait pcmk.resource.utilization pcmk.resource.utilization-set-list-for-resource pcmk.stonith.cleanup pcmk.stonith.cleanup.one-resource pcmk.stonith.cleanup.strict pcmk.stonith.create pcmk.stonith.create.meta pcmk.stonith.create.operations pcmk.stonith.create.operations.onfail-demote pcmk.stonith.create.wait pcmk.stonith.delete pcmk.stonith.enable-disable pcmk.stonith.enable-disable.list pcmk.stonith.enable-disable.wait pcmk.stonith.history.cleanup pcmk.stonith.history.show pcmk.stonith.history.update pcmk.stonith.levels pcmk.stonith.levels.add-remove-devices-list pcmk.stonith.levels.clear pcmk.stonith.levels.node-attr pcmk.stonith.levels.node-regexp pcmk.stonith.levels.verify pcmk.stonith.refresh pcmk.stonith.refresh.one-resource pcmk.stonith.refresh.strict pcmk.stonith.update pcmk.stonith.update.scsi-devices pcmk.stonith.update.scsi-devices.add-remove pcmk.stonith.update.scsi-devices.mpath pcmk.tag pcmk.tag.resources pcs.auth.client pcs.auth.client.cluster pcs.auth.client.token pcs.auth.deauth-client pcs.auth.deauth-server pcs.auth.no-bidirectional pcs.auth.separated-name-and-address pcs.auth.server.token pcs.cfg-in-file.cib pcs.daemon-ssl-cert.set pcs.daemon-ssl-cert.sync-to-local-cluster pcs.disaster-recovery.essentials pcs.reports.severity.deprecation pcs.request-timeout pcs.yes-to-confirm-commands resource-agents.describe resource-agents.list resource-agents.list.detailed resource-agents.ocf.version-1-0 resource-agents.ocf.version-1-1 resource-agents.self-validation sbd sbd.option-timeout-action sbd.shared-block-device status.corosync.membership status.pcmk.query.resource status.pcmk.resources.hide-inactive status.pcmk.resources.id status.pcmk.resources.node status.pcmk.resources.orphaned status.pcmk.wait status.pcmk.xml stonith-agents.describe stonith-agents.list stonith-agents.list.detailed stonith-agents.ocf.version-1-0 stonith-agents.ocf.version-1-1 stonith-agents.self-validation

TASK [fedora.linux_system_roles.ha_cluster : Parse pcs capabilities] ***********
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:148
Monday 29 July 2024  07:42:20 -0400 (0:00:01.934)       0:00:08.101 *********** 
ok: [managed_node1] => {
    "ansible_facts": {
        "__ha_cluster_pcs_capabilities": [
            "booth",
            "booth.enable-authfile.unset",
            "cluster.config.backup-local",
            "cluster.config.restore-cluster",
            "cluster.config.restore-local",
            "cluster.config.uuid",
            "cluster.create",
            "cluster.create.enable",
            "cluster.create.local",
            "cluster.create.no-keys-sync",
            "cluster.create.separated-name-and-address",
            "cluster.create.start",
            "cluster.create.start.wait",
            "cluster.create.transport.knet",
            "cluster.create.transport.udp-udpu",
            "cluster.create.transport.udp-udpu.no-rrp",
            "cluster.destroy",
            "cluster.destroy.all",
            "cluster.report",
            "cluster.verify",
            "corosync.authkey.update",
            "corosync.config.get",
            "corosync.config.get.struct",
            "corosync.config.reload",
            "corosync.config.sync-to-local-cluster",
            "corosync.config.update",
            "corosync.link.add",
            "corosync.link.remove",
            "corosync.link.remove.list",
            "corosync.link.update",
            "corosync.qdevice",
            "corosync.qdevice.model.net",
            "corosync.quorum",
            "corosync.quorum.device",
            "corosync.quorum.device.client.model.net.certificates.local",
            "corosync.quorum.device.heuristics",
            "corosync.quorum.device.model.net",
            "corosync.quorum.device.model.net.options_tls_and_kaptb",
            "corosync.quorum.set-expected-votes-runtime",
            "corosync.quorum.status",
            "corosync.quorum.unblock",
            "corosync.totem.block_unlisted_ips",
            "corosync.uidgid",
            "node.add",
            "node.add.enable",
            "node.add.separated-name-and-address",
            "node.add.start",
            "node.add.start.wait",
            "node.attributes",
            "node.attributes.set-list-for-node",
            "node.confirm-off",
            "node.fence",
            "node.guest",
            "node.kill",
            "node.maintenance",
            "node.maintenance.all",
            "node.maintenance.list",
            "node.maintenance.wait",
            "node.remote",
            "node.remote.onfail-demote",
            "node.remove",
            "node.remove-from-caches",
            "node.remove.list",
            "node.standby",
            "node.standby.all",
            "node.standby.list",
            "node.standby.wait",
            "node.start-stop-enable-disable",
            "node.start-stop-enable-disable.all",
            "node.start-stop-enable-disable.list",
            "node.start-stop-enable-disable.start-wait",
            "node.utilization",
            "node.utilization.set-list-for-node",
            "pcmk.acl.enable-disable",
            "pcmk.acl.group",
            "pcmk.acl.role",
            "pcmk.acl.role.create-with-permissions",
            "pcmk.acl.role.delete-with-users-groups",
            "pcmk.acl.user",
            "pcmk.alert",
            "pcmk.cib.checkpoints",
            "pcmk.cib.checkpoints.diff",
            "pcmk.cib.edit",
            "pcmk.cib.get",
            "pcmk.cib.get.scope",
            "pcmk.cib.roles.promoted-unpromoted",
            "pcmk.cib.set",
            "pcmk.constraint.colocation.set",
            "pcmk.constraint.colocation.set.options",
            "pcmk.constraint.colocation.simple",
            "pcmk.constraint.colocation.simple.options",
            "pcmk.constraint.colocation.simple.options.score",
            "pcmk.constraint.config.output-formats",
            "pcmk.constraint.hide-expired",
            "pcmk.constraint.location.simple",
            "pcmk.constraint.location.simple.options",
            "pcmk.constraint.location.simple.options.score",
            "pcmk.constraint.location.simple.resource-regexp",
            "pcmk.constraint.location.simple.rule",
            "pcmk.constraint.location.simple.rule.node-attr-type-number",
            "pcmk.constraint.location.simple.rule.options",
            "pcmk.constraint.location.simple.rule.rule-add-remove",
            "pcmk.constraint.location.simple.rule.rule-as-one-argument",
            "pcmk.constraint.no-autocorrect",
            "pcmk.constraint.order.set",
            "pcmk.constraint.order.set.options",
            "pcmk.constraint.order.simple",
            "pcmk.constraint.order.simple.options",
            "pcmk.constraint.ticket.set",
            "pcmk.constraint.ticket.set.options",
            "pcmk.constraint.ticket.simple",
            "pcmk.constraint.ticket.simple.constraint-id",
            "pcmk.properties.cluster",
            "pcmk.properties.cluster.config.output-formats",
            "pcmk.properties.cluster.defaults",
            "pcmk.properties.cluster.describe",
            "pcmk.properties.cluster.describe.output-formats",
            "pcmk.properties.cluster.set_properties.iso8601duration",
            "pcmk.properties.operation-defaults",
            "pcmk.properties.operation-defaults.config.output-formats",
            "pcmk.properties.operation-defaults.multiple",
            "pcmk.properties.operation-defaults.rule",
            "pcmk.properties.operation-defaults.rule-rsc-op",
            "pcmk.properties.operation-defaults.rule.hide-expired",
            "pcmk.properties.operation-defaults.rule.node-attr-type-number",
            "pcmk.properties.resource-defaults",
            "pcmk.properties.resource-defaults.config.output-formats",
            "pcmk.properties.resource-defaults.multiple",
            "pcmk.properties.resource-defaults.rule",
            "pcmk.properties.resource-defaults.rule-rsc-op",
            "pcmk.properties.resource-defaults.rule.hide-expired",
            "pcmk.properties.resource-defaults.rule.node-attr-type-number",
            "pcmk.resource.ban-move-clear",
            "pcmk.resource.ban-move-clear.bundles",
            "pcmk.resource.ban-move-clear.clear-expired",
            "pcmk.resource.ban-move-clear.clone",
            "pcmk.resource.bundle",
            "pcmk.resource.bundle.container-docker",
            "pcmk.resource.bundle.container-docker.promoted-max",
            "pcmk.resource.bundle.container-podman",
            "pcmk.resource.bundle.container-podman.promoted-max",
            "pcmk.resource.bundle.container-rkt",
            "pcmk.resource.bundle.container-rkt.promoted-max",
            "pcmk.resource.bundle.reset",
            "pcmk.resource.bundle.wait",
            "pcmk.resource.cleanup",
            "pcmk.resource.cleanup.one-resource",
            "pcmk.resource.cleanup.strict",
            "pcmk.resource.clone",
            "pcmk.resource.clone.custom-id",
            "pcmk.resource.clone.meta-in-create",
            "pcmk.resource.clone.wait",
            "pcmk.resource.config.output-formats",
            "pcmk.resource.create",
            "pcmk.resource.create.clone.custom-id",
            "pcmk.resource.create.group.future",
            "pcmk.resource.create.in-existing-bundle",
            "pcmk.resource.create.meta",
            "pcmk.resource.create.meta.future",
            "pcmk.resource.create.no-master",
            "pcmk.resource.create.operations",
            "pcmk.resource.create.operations.onfail-demote",
            "pcmk.resource.create.promotable",
            "pcmk.resource.create.promotable.custom-id",
            "pcmk.resource.create.wait",
            "pcmk.resource.debug",
            "pcmk.resource.delete",
            "pcmk.resource.disable.safe",
            "pcmk.resource.disable.safe.brief",
            "pcmk.resource.disable.safe.tag",
            "pcmk.resource.disable.simulate",
            "pcmk.resource.disable.simulate.brief",
            "pcmk.resource.disable.simulate.tag",
            "pcmk.resource.enable-disable",
            "pcmk.resource.enable-disable.list",
            "pcmk.resource.enable-disable.tag",
            "pcmk.resource.enable-disable.wait",
            "pcmk.resource.failcount",
            "pcmk.resource.group",
            "pcmk.resource.group.add-remove-list",
            "pcmk.resource.group.wait",
            "pcmk.resource.manage-unmanage",
            "pcmk.resource.manage-unmanage.list",
            "pcmk.resource.manage-unmanage.tag",
            "pcmk.resource.manage-unmanage.with-monitor",
            "pcmk.resource.move.autoclean",
            "pcmk.resource.move.autoclean.default",
            "pcmk.resource.promotable",
            "pcmk.resource.promotable.custom-id",
            "pcmk.resource.promotable.meta-in-create",
            "pcmk.resource.promotable.wait",
            "pcmk.resource.refresh",
            "pcmk.resource.refresh.one-resource",
            "pcmk.resource.refresh.strict",
            "pcmk.resource.relations",
            "pcmk.resource.relocate",
            "pcmk.resource.restart",
            "pcmk.resource.update",
            "pcmk.resource.update-meta",
            "pcmk.resource.update-meta.list",
            "pcmk.resource.update-meta.wait",
            "pcmk.resource.update-operations",
            "pcmk.resource.update-operations.onfail-demote",
            "pcmk.resource.update.meta",
            "pcmk.resource.update.operations",
            "pcmk.resource.update.operations.onfail-demote",
            "pcmk.resource.update.wait",
            "pcmk.resource.utilization",
            "pcmk.resource.utilization-set-list-for-resource",
            "pcmk.stonith.cleanup",
            "pcmk.stonith.cleanup.one-resource",
            "pcmk.stonith.cleanup.strict",
            "pcmk.stonith.create",
            "pcmk.stonith.create.meta",
            "pcmk.stonith.create.operations",
            "pcmk.stonith.create.operations.onfail-demote",
            "pcmk.stonith.create.wait",
            "pcmk.stonith.delete",
            "pcmk.stonith.enable-disable",
            "pcmk.stonith.enable-disable.list",
            "pcmk.stonith.enable-disable.wait",
            "pcmk.stonith.history.cleanup",
            "pcmk.stonith.history.show",
            "pcmk.stonith.history.update",
            "pcmk.stonith.levels",
            "pcmk.stonith.levels.add-remove-devices-list",
            "pcmk.stonith.levels.clear",
            "pcmk.stonith.levels.node-attr",
            "pcmk.stonith.levels.node-regexp",
            "pcmk.stonith.levels.verify",
            "pcmk.stonith.refresh",
            "pcmk.stonith.refresh.one-resource",
            "pcmk.stonith.refresh.strict",
            "pcmk.stonith.update",
            "pcmk.stonith.update.scsi-devices",
            "pcmk.stonith.update.scsi-devices.add-remove",
            "pcmk.stonith.update.scsi-devices.mpath",
            "pcmk.tag",
            "pcmk.tag.resources",
            "pcs.auth.client",
            "pcs.auth.client.cluster",
            "pcs.auth.client.token",
            "pcs.auth.deauth-client",
            "pcs.auth.deauth-server",
            "pcs.auth.no-bidirectional",
            "pcs.auth.separated-name-and-address",
            "pcs.auth.server.token",
            "pcs.cfg-in-file.cib",
            "pcs.daemon-ssl-cert.set",
            "pcs.daemon-ssl-cert.sync-to-local-cluster",
            "pcs.disaster-recovery.essentials",
            "pcs.reports.severity.deprecation",
            "pcs.request-timeout",
            "pcs.yes-to-confirm-commands",
            "resource-agents.describe",
            "resource-agents.list",
            "resource-agents.list.detailed",
            "resource-agents.ocf.version-1-0",
            "resource-agents.ocf.version-1-1",
            "resource-agents.self-validation",
            "sbd",
            "sbd.option-timeout-action",
            "sbd.shared-block-device",
            "status.corosync.membership",
            "status.pcmk.query.resource",
            "status.pcmk.resources.hide-inactive",
            "status.pcmk.resources.id",
            "status.pcmk.resources.node",
            "status.pcmk.resources.orphaned",
            "status.pcmk.wait",
            "status.pcmk.xml",
            "stonith-agents.describe",
            "stonith-agents.list",
            "stonith-agents.list.detailed",
            "stonith-agents.ocf.version-1-0",
            "stonith-agents.ocf.version-1-1",
            "stonith-agents.self-validation"
        ],
        "__ha_cluster_pcsd_capabilities_available": true
    },
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Fetch pcsd capabilities] **********
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:155
Monday 29 July 2024  07:42:20 -0400 (0:00:00.025)       0:00:08.126 *********** 
ok: [managed_node1] => {
    "changed": false,
    "cmd": [
        "pcsd",
        "--version",
        "--full"
    ],
    "delta": "0:00:00.661908",
    "end": "2024-07-29 07:42:21.554117",
    "rc": 0,
    "start": "2024-07-29 07:42:20.892209"
}

STDOUT:

0.12.0a1
booth.get-config booth.set-config booth.set-config.multiple cluster.config.restore-local cluster.config.uuid cluster.create cluster.create.enable cluster.create.no-keys-sync cluster.create.separated-name-and-address cluster.create.start cluster.create.start.wait cluster.create.transport.knet cluster.create.transport.udp-udpu cluster.create.transport.udp-udpu.no-rrp cluster.destroy cluster.destroy.all corosync.config.get corosync.config.set corosync.qdevice.model.net.certificates corosync.quorum.device.client corosync.quorum.device.client.model.net.certificates corosync.quorum.device.client.model.net.certificates.local corosync.quorum.device.client.model.net.certificates.rest-api.v2 corosync.quorum.status corosync.totem.block_unlisted_ips node.add node.add.enable node.add.list node.add.separated-name-and-address node.add.start node.add.start.wait node.attributes node.maintenance node.maintenance.all node.maintenance.list node.maintenance.wait node.remove node.remove-from-caches node.remove.list node.standby node.standby.all node.standby.list node.standby.wait node.start-stop-enable-disable node.start-stop-enable-disable.all node.start-stop-enable-disable.stop-component node.utilization pcmk.acl.group pcmk.acl.role pcmk.acl.role.create-with-permissions pcmk.acl.role.delete-with-users-groups pcmk.acl.role.delete-with-users-groups-implicit pcmk.acl.user pcmk.alert pcmk.cib.get pcmk.cib.remove_elements.constraints pcmk.cib.remove_elements.rest-api.v2 pcmk.cib.roles.promoted-unpromoted pcmk.constraint.colocation.set pcmk.constraint.colocation.set.options pcmk.constraint.colocation.simple pcmk.constraint.config.output-formats pcmk.constraint.location.simple pcmk.constraint.location.simple.rule pcmk.constraint.location.simple.rule.node-attr-type-number pcmk.constraint.location.simple.rule.rest-api.v2 pcmk.constraint.no-autocorrect pcmk.constraint.order.set pcmk.constraint.order.set.options pcmk.constraint.order.simple pcmk.constraint.ticket.set pcmk.constraint.ticket.set.options pcmk.constraint.ticket.simple pcmk.constraint.ticket.simple.constraint-id pcmk.properties.cluster pcmk.properties.cluster.describe pcmk.properties.cluster.describe.rest-api.v2 pcmk.properties.cluster.get_properties.rest-api.v2 pcmk.properties.cluster.set_properties.iso8601duration pcmk.properties.cluster.set_properties.rest-api.v2 pcmk.properties.operation-defaults.config.rest-api.v2 pcmk.properties.resource-defaults.config.rest-api.v2 pcmk.resource.ban-move-clear pcmk.resource.ban-move-clear.bundles pcmk.resource.ban-move-clear.clear-expired pcmk.resource.ban-move-clear.clone pcmk.resource.cleanup.one-resource pcmk.resource.cleanup.strict pcmk.resource.clone pcmk.resource.create pcmk.resource.create.clone.custom-id pcmk.resource.create.meta pcmk.resource.create.no-master pcmk.resource.create.operations pcmk.resource.create.operations.onfail-demote pcmk.resource.create.promotable pcmk.resource.create.promotable.custom-id pcmk.resource.create.wait pcmk.resource.delete pcmk.resource.delete.list pcmk.resource.disable.safe pcmk.resource.disable.simulate pcmk.resource.disable.simulate.brief pcmk.resource.disable.simulate.tag pcmk.resource.enable-disable pcmk.resource.enable-disable.list pcmk.resource.enable-disable.wait pcmk.resource.group pcmk.resource.manage-unmanage pcmk.resource.manage-unmanage.list pcmk.resource.manage-unmanage.tag pcmk.resource.manage-unmanage.with-monitor pcmk.resource.move.autoclean pcmk.resource.promotable pcmk.resource.refresh.one-resource pcmk.resource.refresh.strict pcmk.resource.update pcmk.resource.update-meta pcmk.resource.utilization pcmk.stonith.cleanup.one-resource pcmk.stonith.cleanup.strict pcmk.stonith.create pcmk.stonith.create.meta pcmk.stonith.create.operations pcmk.stonith.create.operations.onfail-demote pcmk.stonith.create.wait pcmk.stonith.delete pcmk.stonith.delete.list pcmk.stonith.enable-disable pcmk.stonith.enable-disable.list pcmk.stonith.enable-disable.wait pcmk.stonith.levels pcmk.stonith.levels.add-remove-devices-list pcmk.stonith.levels.clear pcmk.stonith.levels.node-attr pcmk.stonith.levels.node-regexp pcmk.stonith.levels.verify pcmk.stonith.refresh.one-resource pcmk.stonith.refresh.strict pcmk.stonith.scsi-unfence-node-mpath pcmk.stonith.scsi-unfence-node-v2 pcmk.stonith.update pcs.auth.export-cluster-known-hosts pcs.auth.known-host-change pcs.auth.no-bidirectional pcs.auth.separated-name-and-address pcs.auth.server pcs.automatic-pcs-configs-sync pcs.daemon-ssl-cert.set pcs.permissions pcs.rest-api.v1.1 pcs.rest-api.v2 resource-agents.describe resource-agents.list resource-agents.list.detailed resource-agents.ocf.version-1-0 resource-agents.ocf.version-1-1 resource-agents.self-validation resource-agents.simplified-query-api sbd sbd-node sbd-node.shared-block-device sbd.option-timeout-action status.pcmk.local-node status.pcmk.resources.rest-api.v2 status.pcmk.wait stonith-agents.describe stonith-agents.list stonith-agents.list.detailed stonith-agents.ocf.version-1-0 stonith-agents.ocf.version-1-1 stonith-agents.self-validation

TASK [fedora.linux_system_roles.ha_cluster : Parse pcsd capabilities] **********
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:163
Monday 29 July 2024  07:42:21 -0400 (0:00:00.991)       0:00:09.118 *********** 
ok: [managed_node1] => {
    "ansible_facts": {
        "__ha_cluster_pcsd_capabilities": [
            "booth.get-config",
            "booth.set-config",
            "booth.set-config.multiple",
            "cluster.config.restore-local",
            "cluster.config.uuid",
            "cluster.create",
            "cluster.create.enable",
            "cluster.create.no-keys-sync",
            "cluster.create.separated-name-and-address",
            "cluster.create.start",
            "cluster.create.start.wait",
            "cluster.create.transport.knet",
            "cluster.create.transport.udp-udpu",
            "cluster.create.transport.udp-udpu.no-rrp",
            "cluster.destroy",
            "cluster.destroy.all",
            "corosync.config.get",
            "corosync.config.set",
            "corosync.qdevice.model.net.certificates",
            "corosync.quorum.device.client",
            "corosync.quorum.device.client.model.net.certificates",
            "corosync.quorum.device.client.model.net.certificates.local",
            "corosync.quorum.device.client.model.net.certificates.rest-api.v2",
            "corosync.quorum.status",
            "corosync.totem.block_unlisted_ips",
            "node.add",
            "node.add.enable",
            "node.add.list",
            "node.add.separated-name-and-address",
            "node.add.start",
            "node.add.start.wait",
            "node.attributes",
            "node.maintenance",
            "node.maintenance.all",
            "node.maintenance.list",
            "node.maintenance.wait",
            "node.remove",
            "node.remove-from-caches",
            "node.remove.list",
            "node.standby",
            "node.standby.all",
            "node.standby.list",
            "node.standby.wait",
            "node.start-stop-enable-disable",
            "node.start-stop-enable-disable.all",
            "node.start-stop-enable-disable.stop-component",
            "node.utilization",
            "pcmk.acl.group",
            "pcmk.acl.role",
            "pcmk.acl.role.create-with-permissions",
            "pcmk.acl.role.delete-with-users-groups",
            "pcmk.acl.role.delete-with-users-groups-implicit",
            "pcmk.acl.user",
            "pcmk.alert",
            "pcmk.cib.get",
            "pcmk.cib.remove_elements.constraints",
            "pcmk.cib.remove_elements.rest-api.v2",
            "pcmk.cib.roles.promoted-unpromoted",
            "pcmk.constraint.colocation.set",
            "pcmk.constraint.colocation.set.options",
            "pcmk.constraint.colocation.simple",
            "pcmk.constraint.config.output-formats",
            "pcmk.constraint.location.simple",
            "pcmk.constraint.location.simple.rule",
            "pcmk.constraint.location.simple.rule.node-attr-type-number",
            "pcmk.constraint.location.simple.rule.rest-api.v2",
            "pcmk.constraint.no-autocorrect",
            "pcmk.constraint.order.set",
            "pcmk.constraint.order.set.options",
            "pcmk.constraint.order.simple",
            "pcmk.constraint.ticket.set",
            "pcmk.constraint.ticket.set.options",
            "pcmk.constraint.ticket.simple",
            "pcmk.constraint.ticket.simple.constraint-id",
            "pcmk.properties.cluster",
            "pcmk.properties.cluster.describe",
            "pcmk.properties.cluster.describe.rest-api.v2",
            "pcmk.properties.cluster.get_properties.rest-api.v2",
            "pcmk.properties.cluster.set_properties.iso8601duration",
            "pcmk.properties.cluster.set_properties.rest-api.v2",
            "pcmk.properties.operation-defaults.config.rest-api.v2",
            "pcmk.properties.resource-defaults.config.rest-api.v2",
            "pcmk.resource.ban-move-clear",
            "pcmk.resource.ban-move-clear.bundles",
            "pcmk.resource.ban-move-clear.clear-expired",
            "pcmk.resource.ban-move-clear.clone",
            "pcmk.resource.cleanup.one-resource",
            "pcmk.resource.cleanup.strict",
            "pcmk.resource.clone",
            "pcmk.resource.create",
            "pcmk.resource.create.clone.custom-id",
            "pcmk.resource.create.meta",
            "pcmk.resource.create.no-master",
            "pcmk.resource.create.operations",
            "pcmk.resource.create.operations.onfail-demote",
            "pcmk.resource.create.promotable",
            "pcmk.resource.create.promotable.custom-id",
            "pcmk.resource.create.wait",
            "pcmk.resource.delete",
            "pcmk.resource.delete.list",
            "pcmk.resource.disable.safe",
            "pcmk.resource.disable.simulate",
            "pcmk.resource.disable.simulate.brief",
            "pcmk.resource.disable.simulate.tag",
            "pcmk.resource.enable-disable",
            "pcmk.resource.enable-disable.list",
            "pcmk.resource.enable-disable.wait",
            "pcmk.resource.group",
            "pcmk.resource.manage-unmanage",
            "pcmk.resource.manage-unmanage.list",
            "pcmk.resource.manage-unmanage.tag",
            "pcmk.resource.manage-unmanage.with-monitor",
            "pcmk.resource.move.autoclean",
            "pcmk.resource.promotable",
            "pcmk.resource.refresh.one-resource",
            "pcmk.resource.refresh.strict",
            "pcmk.resource.update",
            "pcmk.resource.update-meta",
            "pcmk.resource.utilization",
            "pcmk.stonith.cleanup.one-resource",
            "pcmk.stonith.cleanup.strict",
            "pcmk.stonith.create",
            "pcmk.stonith.create.meta",
            "pcmk.stonith.create.operations",
            "pcmk.stonith.create.operations.onfail-demote",
            "pcmk.stonith.create.wait",
            "pcmk.stonith.delete",
            "pcmk.stonith.delete.list",
            "pcmk.stonith.enable-disable",
            "pcmk.stonith.enable-disable.list",
            "pcmk.stonith.enable-disable.wait",
            "pcmk.stonith.levels",
            "pcmk.stonith.levels.add-remove-devices-list",
            "pcmk.stonith.levels.clear",
            "pcmk.stonith.levels.node-attr",
            "pcmk.stonith.levels.node-regexp",
            "pcmk.stonith.levels.verify",
            "pcmk.stonith.refresh.one-resource",
            "pcmk.stonith.refresh.strict",
            "pcmk.stonith.scsi-unfence-node-mpath",
            "pcmk.stonith.scsi-unfence-node-v2",
            "pcmk.stonith.update",
            "pcs.auth.export-cluster-known-hosts",
            "pcs.auth.known-host-change",
            "pcs.auth.no-bidirectional",
            "pcs.auth.separated-name-and-address",
            "pcs.auth.server",
            "pcs.automatic-pcs-configs-sync",
            "pcs.daemon-ssl-cert.set",
            "pcs.permissions",
            "pcs.rest-api.v1.1",
            "pcs.rest-api.v2",
            "resource-agents.describe",
            "resource-agents.list",
            "resource-agents.list.detailed",
            "resource-agents.ocf.version-1-0",
            "resource-agents.ocf.version-1-1",
            "resource-agents.self-validation",
            "resource-agents.simplified-query-api",
            "sbd",
            "sbd-node",
            "sbd-node.shared-block-device",
            "sbd.option-timeout-action",
            "status.pcmk.local-node",
            "status.pcmk.resources.rest-api.v2",
            "status.pcmk.wait",
            "stonith-agents.describe",
            "stonith-agents.list",
            "stonith-agents.list.detailed",
            "stonith-agents.ocf.version-1-0",
            "stonith-agents.ocf.version-1-1",
            "stonith-agents.self-validation"
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.ha_cluster : Fail if pcs is to old to configure resources and operations defaults] ***
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:172
Monday 29 July 2024  07:42:21 -0400 (0:00:00.024)       0:00:09.142 *********** 
skipping: [managed_node1] => {
    "changed": false,
    "false_condition": "( ha_cluster_resource_defaults and not 'pcmk.properties.resource-defaults.multiple' in __ha_cluster_pcs_capabilities ) or ( ha_cluster_resource_operation_defaults and not 'pcmk.properties.operation-defaults.multiple' in __ha_cluster_pcs_capabilities )",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.ha_cluster : Set hacluster password] ***********
task path: /tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:22
Monday 29 July 2024  07:42:21 -0400 (0:00:00.014)       0:00:09.157 *********** 
fatal: [managed_node1]: FAILED! => {}

MSG:

Unable to encrypt nor hash, passlib must be installed. No module named 'passlib'. Unable to encrypt nor hash, passlib must be installed. No module named 'passlib'

PLAY RECAP *********************************************************************
managed_node1              : ok=28   changed=2    unreachable=0    failed=1    skipped=16   rescued=0    ignored=0   

Monday 29 July 2024  07:42:21 -0400 (0:00:00.055)       0:00:09.213 *********** 
=============================================================================== 
fedora.linux_system_roles.ha_cluster : Fetch pcs capabilities ----------- 1.93s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:141 
fedora.linux_system_roles.ha_cluster : Install role essential packages --- 1.50s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:11 
fedora.linux_system_roles.ha_cluster : Make sure qnetd is not installed --- 1.40s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_cleanup_qnetd.yml:9 
Gathering Facts --------------------------------------------------------- 1.08s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_qnetd_disabled.yml:9 
fedora.linux_system_roles.ha_cluster : Fetch pcsd capabilities ---------- 0.99s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:155 
fedora.linux_system_roles.ha_cluster : List active CentOS repositories --- 0.59s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-repositories/CentOS.yml:3 
fedora.linux_system_roles.ha_cluster : Make sure qnetd config files are not present --- 0.43s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_cleanup_qnetd.yml:16 
fedora.linux_system_roles.ha_cluster : Check if system is ostree -------- 0.40s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/test_setup.yml:22 
fedora.linux_system_roles.ha_cluster : Set hacluster password ----------- 0.06s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:22 
Run HA Cluster role ----------------------------------------------------- 0.04s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_qnetd_disabled.yml:31 
fedora.linux_system_roles.ha_cluster : Fail if passwords are not specified --- 0.04s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:43 
fedora.linux_system_roles.ha_cluster : Set platform/version specific variables --- 0.04s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/set_vars.yml:19 
fedora.linux_system_roles.ha_cluster : Find platform/version specific tasks to enable repositories --- 0.04s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-package-repositories.yml:3 
fedora.linux_system_roles.ha_cluster : Check and prepare role variables --- 0.04s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/main.yml:17 
Include vault variables ------------------------------------------------- 0.03s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_qnetd_disabled.yml:5 
fedora.linux_system_roles.ha_cluster : Run platform/version specific tasks to enable repositories --- 0.03s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/enable-package-repositories.yml:21 
fedora.linux_system_roles.ha_cluster : Extract qdevice settings --------- 0.03s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:101 
Set up test environment ------------------------------------------------- 0.03s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/tests/ha_cluster/tests_qnetd_disabled.yml:21 
fedora.linux_system_roles.ha_cluster : Parse pcs capabilities ----------- 0.03s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:148 
fedora.linux_system_roles.ha_cluster : Extract node options ------------- 0.02s
/tmp/tmp.JVrTHQ6u36/ansible_collections/fedora/linux_system_roles/roles/ha_cluster/tasks/shell_pcs/check-and-prepare-role-variables.yml:30