ansible-playbook [core 2.17.13] config file = None configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/local/lib/python3.12/site-packages/ansible ansible collection location = /tmp/collections-sHE executable location = /usr/local/bin/ansible-playbook python version = 3.12.11 (main, Aug 14 2025, 00:00:00) [GCC 11.5.0 20240719 (Red Hat 11.5.0-11)] (/usr/bin/python3.12) jinja version = 3.1.6 libyaml = True No config file found; using defaults running playbook inside collection fedora.linux_system_roles Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'json', as we already have a stdout callback. Skipping callback 'jsonl', as we already have a stdout callback. Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_default.yml **************************************************** 1 plays in /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/tests/hpc/tests_default.yml PLAY [Ensure that the role runs with default parameters] *********************** TASK [fedora.linux_system_roles.hpc : Set platform/version specific variables] *** task path: /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:3 Monday 01 September 2025 10:45:31 -0400 (0:00:00.021) 0:00:00.021 ****** included: /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/set_vars.yml for managed-node1 TASK [fedora.linux_system_roles.hpc : Ensure ansible_facts used by role] ******* task path: /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/set_vars.yml:2 Monday 01 September 2025 10:45:31 -0400 (0:00:00.026) 0:00:00.048 ****** [WARNING]: Platform linux on host managed-node1 is using the discovered Python interpreter at /usr/bin/python3.9, but future installation of another Python interpreter could change the meaning of that path. See https://docs.ansible.com/ansible- core/2.17/reference_appendices/interpreter_discovery.html for more information. ok: [managed-node1] TASK [fedora.linux_system_roles.hpc : Check if system is ostree] *************** task path: /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/set_vars.yml:10 Monday 01 September 2025 10:45:32 -0400 (0:00:00.751) 0:00:00.800 ****** ok: [managed-node1] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.hpc : Set flag to indicate system is ostree] *** task path: /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/set_vars.yml:15 Monday 01 September 2025 10:45:32 -0400 (0:00:00.427) 0:00:01.227 ****** ok: [managed-node1] => { "ansible_facts": { "__hpc_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.hpc : Set platform/version specific variables] *** task path: /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/set_vars.yml:19 Monday 01 September 2025 10:45:32 -0400 (0:00:00.019) 0:00:01.247 ****** skipping: [managed-node1] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node1] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node1] => (item=CentOS_9.yml) => { "ansible_facts": { "__template_packages": [], "__template_services": [] }, "ansible_included_var_files": [ "/tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/vars/CentOS_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_9.yml" } ok: [managed-node1] => (item=CentOS_9.yml) => { "ansible_facts": { "__template_packages": [], "__template_services": [] }, "ansible_included_var_files": [ "/tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/vars/CentOS_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_9.yml" } TASK [fedora.linux_system_roles.hpc : Deploy the GPG key for RHEL EPEL repository] *** task path: /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:6 Monday 01 September 2025 10:45:32 -0400 (0:00:00.035) 0:00:01.283 ****** ok: [managed-node1] => { "changed": false } TASK [fedora.linux_system_roles.hpc : Install EPEL release package] ************ task path: /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:12 Monday 01 September 2025 10:45:33 -0400 (0:00:00.643) 0:00:01.926 ****** ok: [managed-node1] => { "changed": false, "rc": 0, "results": [ "Installed /root/.ansible/tmp/ansible-tmp-1756737933.3425984-8996-44713145852643/epel-release-latest-9.noarch2weebiba.rpm" ] } MSG: Nothing to do TASK [fedora.linux_system_roles.hpc : Deploy the GPG key for NVIDIA repositories] *** task path: /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:18 Monday 01 September 2025 10:45:35 -0400 (0:00:02.020) 0:00:03.947 ****** changed: [managed-node1] => { "changed": true } TASK [fedora.linux_system_roles.hpc : Configure the NVIDIA CUDA repository] **** task path: /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:23 Monday 01 September 2025 10:45:35 -0400 (0:00:00.476) 0:00:04.424 ****** redirecting (type: action) ansible.builtin.yum to ansible.builtin.dnf changed: [managed-node1] => { "changed": true, "repo": "nvidia-cuda", "state": "present" } TASK [fedora.linux_system_roles.hpc : Install lvm2 to get lvs command] ********* task path: /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:33 Monday 01 September 2025 10:45:36 -0400 (0:00:00.426) 0:00:04.851 ****** skipping: [managed-node1] => { "changed": false, "false_condition": "hpc_manage_storage", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.hpc : Get current LV size of rootlv] *********** task path: /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:39 Monday 01 September 2025 10:45:36 -0400 (0:00:00.014) 0:00:04.865 ****** skipping: [managed-node1] => { "changed": false, "false_condition": "hpc_manage_storage", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.hpc : Get current LV size of usrlv] ************ task path: /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:46 Monday 01 September 2025 10:45:36 -0400 (0:00:00.011) 0:00:04.876 ****** skipping: [managed-node1] => { "changed": false, "false_condition": "hpc_manage_storage", "skip_reason": "Conditional result was False" } TASK [Configure storage] ******************************************************* task path: /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:53 Monday 01 September 2025 10:45:36 -0400 (0:00:00.009) 0:00:04.885 ****** skipping: [managed-node1] => { "changed": false, "false_condition": "hpc_manage_storage", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.hpc : Update RHUI packages from Microsoft repositories] *** task path: /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:84 Monday 01 September 2025 10:45:36 -0400 (0:00:00.009) 0:00:04.895 ****** fatal: [managed-node1]: FAILED! => {} MSG: The conditional check 'ansible_system_vendor == "Microsoft Corporation"' failed. The error was: error while evaluating conditional (ansible_system_vendor == "Microsoft Corporation"): 'ansible_system_vendor' is undefined The error appears to be in '/tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml': line 84, column 3, but may be elsewhere in the file depending on the exact syntax problem. The offending line appears to be: # This is only needed for MicrosoftAzure VMs - name: Update RHUI packages from Microsoft repositories ^ here PLAY RECAP ********************************************************************* managed-node1 : ok=9 changed=2 unreachable=0 failed=1 skipped=4 rescued=0 ignored=0 SYSTEM ROLES ERRORS BEGIN v1 [ { "ansible_version": "2.17.13", "end_time": "2025-09-01T14:45:36.260082+00:00Z", "host": "managed-node1", "message": "The conditional check 'ansible_system_vendor == \"Microsoft Corporation\"' failed. The error was: error while evaluating conditional (ansible_system_vendor == \"Microsoft Corporation\"): 'ansible_system_vendor' is undefined\n\nThe error appears to be in '/tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml': line 84, column 3, but may\nbe elsewhere in the file depending on the exact syntax problem.\n\nThe offending line appears to be:\n\n# This is only needed for MicrosoftAzure VMs\n- name: Update RHUI packages from Microsoft repositories\n ^ here\n", "start_time": "2025-09-01T14:45:36.252160+00:00Z", "task_name": "Update RHUI packages from Microsoft repositories", "task_path": "/tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:84" } ] SYSTEM ROLES ERRORS END v1 TASKS RECAP ******************************************************************** Monday 01 September 2025 10:45:36 -0400 (0:00:00.009) 0:00:04.904 ****** =============================================================================== fedora.linux_system_roles.hpc : Install EPEL release package ------------ 2.02s /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:12 fedora.linux_system_roles.hpc : Ensure ansible_facts used by role ------- 0.75s /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/set_vars.yml:2 fedora.linux_system_roles.hpc : Deploy the GPG key for RHEL EPEL repository --- 0.64s /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:6 fedora.linux_system_roles.hpc : Deploy the GPG key for NVIDIA repositories --- 0.48s /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:18 fedora.linux_system_roles.hpc : Check if system is ostree --------------- 0.43s /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/set_vars.yml:10 fedora.linux_system_roles.hpc : Configure the NVIDIA CUDA repository ---- 0.43s /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:23 fedora.linux_system_roles.hpc : Set platform/version specific variables --- 0.04s /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/set_vars.yml:19 fedora.linux_system_roles.hpc : Set platform/version specific variables --- 0.03s /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:3 fedora.linux_system_roles.hpc : Set flag to indicate system is ostree --- 0.02s /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/set_vars.yml:15 fedora.linux_system_roles.hpc : Install lvm2 to get lvs command --------- 0.01s /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:33 fedora.linux_system_roles.hpc : Get current LV size of rootlv ----------- 0.01s /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:39 Configure storage ------------------------------------------------------- 0.01s /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:53 fedora.linux_system_roles.hpc : Get current LV size of usrlv ------------ 0.01s /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:46 fedora.linux_system_roles.hpc : Update RHUI packages from Microsoft repositories --- 0.01s /tmp/collections-sHE/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:84 Sep 01 10:45:30 managed-node1 python3.9[8828]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Sep 01 10:45:30 managed-node1 sshd[8881]: Accepted publickey for root from 10.31.45.122 port 57852 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Sep 01 10:45:30 managed-node1 systemd-logind[607]: New session 13 of user root. ░░ Subject: A new session 13 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 13 has been created for the user root. ░░ ░░ The leading process of the session is 8881. Sep 01 10:45:30 managed-node1 systemd[1]: Started Session 13 of User root. ░░ Subject: A start job for unit session-13.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-13.scope has finished successfully. ░░ ░░ The job identifier is 1851. Sep 01 10:45:30 managed-node1 sshd[8881]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Sep 01 10:45:30 managed-node1 sshd[8884]: Received disconnect from 10.31.45.122 port 57852:11: disconnected by user Sep 01 10:45:30 managed-node1 sshd[8884]: Disconnected from user root 10.31.45.122 port 57852 Sep 01 10:45:30 managed-node1 sshd[8881]: pam_unix(sshd:session): session closed for user root Sep 01 10:45:30 managed-node1 systemd[1]: session-13.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-13.scope has successfully entered the 'dead' state. Sep 01 10:45:30 managed-node1 systemd-logind[607]: Session 13 logged out. Waiting for processes to exit. Sep 01 10:45:30 managed-node1 systemd-logind[607]: Removed session 13. ░░ Subject: Session 13 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 13 has been terminated. Sep 01 10:45:32 managed-node1 python3.9[9082]: ansible-setup Invoked with gather_subset=['!all', '!min', 'distribution', 'distribution_major_version', 'distribution_version', 'os_family'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Sep 01 10:45:32 managed-node1 python3.9[9233]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Sep 01 10:45:33 managed-node1 python3.9[9382]: ansible-rpm_key Invoked with key=https://dl.fedoraproject.org/pub/epel/RPM-GPG-KEY-EPEL-9 state=present validate_certs=True fingerprint=None Sep 01 10:45:33 managed-node1 python3.9[9536]: ansible-ansible.legacy.setup Invoked with filter=['ansible_pkg_mgr'] gather_subset=['!all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Sep 01 10:45:34 managed-node1 python3.9[9613]: ansible-ansible.legacy.dnf Invoked with name=['https://dl.fedoraproject.org/pub/epel/epel-release-latest-9.noarch.rpm'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Sep 01 10:45:35 managed-node1 python3.9[9763]: ansible-rpm_key Invoked with key=https://developer.download.nvidia.com/compute/cuda/repos/rhel9/x86_64/D42D0685.pub state=present validate_certs=True fingerprint=None Sep 01 10:45:36 managed-node1 python3.9[9918]: ansible-yum_repository Invoked with name=nvidia-cuda description=NVIDIA CUDA repository baseurl=['https://developer.download.nvidia.com/compute/cuda/repos/rhel9/x86_64'] gpgcheck=True reposdir=/etc/yum.repos.d state=present unsafe_writes=False bandwidth=None cost=None deltarpm_metadata_percentage=None deltarpm_percentage=None enabled=None enablegroups=None exclude=None failovermethod=None file=None gpgcakey=None gpgkey=None module_hotfixes=None http_caching=None include=None includepkgs=None ip_resolve=None keepalive=None keepcache=None metadata_expire=None metadata_expire_filter=None metalink=None mirrorlist=None mirrorlist_expire=None password=NOT_LOGGING_PARAMETER priority=None protect=None proxy=None proxy_password=NOT_LOGGING_PARAMETER proxy_username=None repo_gpgcheck=None retries=None s3_enabled=None skip_if_unavailable=None sslcacert=None ssl_check_cert_permissions=None sslclientcert=None sslclientkey=None sslverify=None throttle=None timeout=None ui_repoid_vars=None username=None async=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Sep 01 10:45:36 managed-node1 sshd[9943]: Accepted publickey for root from 10.31.45.122 port 59724 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Sep 01 10:45:36 managed-node1 systemd-logind[607]: New session 14 of user root. ░░ Subject: A new session 14 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 14 has been created for the user root. ░░ ░░ The leading process of the session is 9943. Sep 01 10:45:36 managed-node1 systemd[1]: Started Session 14 of User root. ░░ Subject: A start job for unit session-14.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-14.scope has finished successfully. ░░ ░░ The job identifier is 1920. Sep 01 10:45:36 managed-node1 sshd[9943]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Sep 01 10:45:36 managed-node1 sshd[9946]: Received disconnect from 10.31.45.122 port 59724:11: disconnected by user Sep 01 10:45:36 managed-node1 sshd[9946]: Disconnected from user root 10.31.45.122 port 59724 Sep 01 10:45:36 managed-node1 sshd[9943]: pam_unix(sshd:session): session closed for user root Sep 01 10:45:36 managed-node1 systemd[1]: session-14.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-14.scope has successfully entered the 'dead' state. Sep 01 10:45:36 managed-node1 systemd-logind[607]: Session 14 logged out. Waiting for processes to exit. Sep 01 10:45:36 managed-node1 systemd-logind[607]: Removed session 14. ░░ Subject: Session 14 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 14 has been terminated. Sep 01 10:45:36 managed-node1 sshd[9971]: Accepted publickey for root from 10.31.45.122 port 59738 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Sep 01 10:45:36 managed-node1 systemd-logind[607]: New session 15 of user root. ░░ Subject: A new session 15 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 15 has been created for the user root. ░░ ░░ The leading process of the session is 9971. Sep 01 10:45:36 managed-node1 systemd[1]: Started Session 15 of User root. ░░ Subject: A start job for unit session-15.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-15.scope has finished successfully. ░░ ░░ The job identifier is 1989. Sep 01 10:45:36 managed-node1 sshd[9971]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0)