[WARNING]: Collection infra.leapp does not support Ansible version 2.14.18 [WARNING]: running playbook inside collection infra.leapp ansible-playbook [core 2.14.18] config file = /etc/ansible/ansible.cfg configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.9/site-packages/ansible ansible collection location = /root/.ansible/collections:/usr/share/ansible/collections executable location = /usr/bin/ansible-playbook python version = 3.9.25 (main, Mar 9 2026, 00:00:00) [GCC 11.5.0 20240719 (Red Hat 11.5.0-14)] (/usr/bin/python3) jinja version = 3.1.2 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_default.yml **************************************************** 1 plays in /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tests/tests_default.yml PLAY [Test] ******************************************************************** TASK [Gathering Facts] ********************************************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tests/tests_default.yml:2 ok: [managed-node01] TASK [Test | Run role upgrade] ************************************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tests/tests_default.yml:10 TASK [Ensure ansible_facts used by role] *************************************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tasks/main.yml:2 TASK [infra.leapp.common : set_facts | Ensure ansible_facts used by role] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/set_facts.yml:10 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.upgrade : Lock timestamped variables] ************************ task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tasks/main.yml:13 ok: [managed-node01] => {"ansible_facts": {"__leapp_timestamp": "2026-03-25_14-53-32"}, "changed": false} TASK [Initialize lock, logging, and common vars] ******************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tasks/main.yml:17 TASK [infra.leapp.common : init_leapp_log | Ensure ansible_facts used by role] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:9 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/set_facts.yml for managed-node01 TASK [infra.leapp.common : set_facts | Ensure ansible_facts used by role] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/set_facts.yml:10 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : init_leapp_log | Ensure that log directory exists] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:12 changed: [managed-node01] => {"changed": true, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/var/log/leapp", "secontext": "system_u:object_r:var_log_t:s0", "size": 69, "state": "directory", "uid": 0} TASK [infra.leapp.common : init_leapp_log | Check for existing log file] ******* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:20 ok: [managed-node01] => {"changed": false, "stat": {"exists": false}} TASK [infra.leapp.common : init_leapp_log | Fail if log file already exists] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:25 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : init_leapp_log | Create new log file] *************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:34 changed: [managed-node01] => {"changed": true, "checksum": "03a2c14d65d2bfbf3d9d68d5adec03ef0e32bcd3", "dest": "/var/log/leapp/ansible_leapp_upgrade.log", "gid": 0, "group": "root", "md5sum": "6b1be3262f18368c7c7740d7391c77ea", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:var_log_t:s0", "size": 61, "src": "/root/.ansible/tmp/ansible-tmp-1774450413.51255-7567-85934667139363/source", "state": "file", "uid": 0} TASK [infra.leapp.common : init_leapp_log | /etc/ansible/facts.d directory exists] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:44 ok: [managed-node01] => {"changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/ansible/facts.d", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 56, "state": "directory", "uid": 0} TASK [infra.leapp.common : init_leapp_log | Capture current ansible_facts for validation after upgrade] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:52 changed: [managed-node01] => (item=/etc/ansible/facts.d/pre_ipu.fact) => {"ansible_loop_var": "item", "changed": true, "checksum": "b5862513113f5e615de2d88beb7c33e80d0a4fad", "dest": "/etc/ansible/facts.d/pre_ipu.fact", "gid": 0, "group": "root", "item": "/etc/ansible/facts.d/pre_ipu.fact", "md5sum": "d80feb6e94e862ac0e8ac054e2f67e5e", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 13838, "src": "/root/.ansible/tmp/ansible-tmp-1774450414.6085193-7595-77698119217234/source", "state": "file", "uid": 0} changed: [managed-node01] => (item=/var/log/leapp/ansible_leapp_upgrade.log) => {"ansible_loop_var": "item", "changed": true, "checksum": "b5862513113f5e615de2d88beb7c33e80d0a4fad", "dest": "/var/log/leapp/ansible_leapp_upgrade.log", "gid": 0, "group": "root", "item": "/var/log/leapp/ansible_leapp_upgrade.log", "md5sum": "d80feb6e94e862ac0e8ac054e2f67e5e", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:var_log_t:s0", "size": 13838, "src": "/root/.ansible/tmp/ansible-tmp-1774450415.232057-7595-116560572157272/source", "state": "file", "uid": 0} TASK [infra.leapp.common : init_leapp_log | Capture a list of non-rhel versioned packages] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:63 ok: [managed-node01] => {"changed": false, "cmd": "set -o pipefail; export PATH=$PATH; rpm -qa | grep -ve '[\\.|+]el8' | grep -vE '^(gpg-pubkey|libmodulemd|katello-ca-consumer)' | sort", "delta": "0:00:00.854549", "end": "2026-03-25 10:53:37.074106", "failed_when_result": false, "msg": "non-zero return code", "rc": 1, "start": "2026-03-25 10:53:36.219557", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [infra.leapp.common : init_leapp_log | Create fact with the non-rhel versioned packages list] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:77 ok: [managed-node01] => {"ansible_facts": {"non_rhel_packages": []}, "changed": false} TASK [infra.leapp.common : init_leapp_log | Capture the list of non-rhel versioned packages in a separate fact file] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:81 ok: [managed-node01] => (item=/etc/ansible/facts.d/non_rhel_packages.fact) => {"ansible_loop_var": "item", "changed": false, "checksum": "97d170e1550eee4afc0af065b78cda302a97674c", "dest": "/etc/ansible/facts.d/non_rhel_packages.fact", "gid": 0, "group": "root", "item": "/etc/ansible/facts.d/non_rhel_packages.fact", "mode": "0644", "owner": "root", "path": "/etc/ansible/facts.d/non_rhel_packages.fact", "secontext": "system_u:object_r:etc_t:s0", "size": 2, "state": "file", "uid": 0} changed: [managed-node01] => (item=/var/log/leapp/ansible_leapp_upgrade.log) => {"ansible_loop_var": "item", "changed": true, "checksum": "97d170e1550eee4afc0af065b78cda302a97674c", "dest": "/var/log/leapp/ansible_leapp_upgrade.log", "gid": 0, "group": "root", "item": "/var/log/leapp/ansible_leapp_upgrade.log", "md5sum": "d751713988987e9331980363e24189ce", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:var_log_t:s0", "size": 2, "src": "/root/.ansible/tmp/ansible-tmp-1774450417.7449992-7645-96450897488858/source", "state": "file", "uid": 0} TASK [infra.leapp.upgrade : Include tasks for leapp upgrade] ******************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tasks/main.yml:30 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tasks/leapp-upgrade.yml for managed-node01 TASK [leapp-upgrade | Run parse_leapp_report to check for inhibitors] ********** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tasks/leapp-upgrade.yml:2 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.upgrade : leapp-upgrade | Verify no inhibitor results found during preupgrade] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tasks/leapp-upgrade.yml:8 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [leapp-upgrade | Include pre_upgrade_update.yml] ************************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tasks/leapp-upgrade.yml:18 TASK [infra.leapp.common : pre_upgrade_update | Ensure ansible_facts used by role] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/pre_upgrade_update.yml:3 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/set_facts.yml for managed-node01 TASK [infra.leapp.common : set_facts | Ensure ansible_facts used by role] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/set_facts.yml:10 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : pre_upgrade_update | Register with Satellite activation key] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/pre_upgrade_update.yml:6 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [pre_upgrade_update | Include custom_local_repos for local_repos_pre_leapp] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/pre_upgrade_update.yml:13 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : pre_upgrade_update | Install leapp packages] ******** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/pre_upgrade_update.yml:21 ok: [managed-node01] => {"changed": false, "msg": "Nothing to do", "rc": 0, "results": []} TASK [infra.leapp.common : pre_upgrade_update | Get package version lock entries] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/pre_upgrade_update.yml:33 ok: [managed-node01] => {"changed": false, "cmd": ["dnf", "versionlock", "list"], "delta": "0:00:00.357694", "end": "2026-03-25 10:53:43.749686", "failed_when_result": false, "msg": "non-zero return code", "rc": 1, "start": "2026-03-25 10:53:43.391992", "stderr": "No such command: versionlock. Please use /usr/bin/dnf --help\nIt could be a DNF plugin command, try: \"dnf install 'dnf-command(versionlock)'\"", "stderr_lines": ["No such command: versionlock. Please use /usr/bin/dnf --help", "It could be a DNF plugin command, try: \"dnf install 'dnf-command(versionlock)'\""], "stdout": "Updating Subscription Management repositories.\nUnable to read consumer identity\n\nThis system is not registered with an entitlement server. You can use subscription-manager to register.", "stdout_lines": ["Updating Subscription Management repositories.", "Unable to read consumer identity", "", "This system is not registered with an entitlement server. You can use subscription-manager to register."]} TASK [infra.leapp.common : pre_upgrade_update | Remove all package version locks] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/pre_upgrade_update.yml:40 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : pre_upgrade_update | Include update-and-reboot.yml] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/pre_upgrade_update.yml:48 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/update-and-reboot.yml for managed-node01 TASK [infra.leapp.common : update-and-reboot | Ensure ansible_facts used by role] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/update-and-reboot.yml:2 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/set_facts.yml for managed-node01 TASK [infra.leapp.common : set_facts | Ensure ansible_facts used by role] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/set_facts.yml:10 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : update-and-reboot | Ensure all updates are applied] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/update-and-reboot.yml:5 ASYNC OK on managed-node01: jid=j303642143353.16025 changed: [managed-node01] => {"ansible_job_id": "j303642143353.16025", "changed": true, "finished": 1, "msg": "", "rc": 0, "results": ["Installed: libatasmart-0.19-14.el8.x86_64", "Installed: libblockdev-2.28-7.el8_10.x86_64", "Installed: libblockdev-crypto-2.28-7.el8_10.x86_64", "Installed: libkadm5-1.18.2-33.el8_10.x86_64", "Installed: platform-python-3.6.8-74.el8_10.x86_64", "Installed: libblockdev-fs-2.28-7.el8_10.x86_64", "Installed: libsmbios-2.4.1-2.el8.x86_64", "Installed: libblockdev-loop-2.28-7.el8_10.x86_64", "Installed: bubblewrap-0.4.0-2.el8_10.x86_64", "Installed: libblockdev-mdraid-2.28-7.el8_10.x86_64", "Installed: libblockdev-part-2.28-7.el8_10.x86_64", "Installed: libblockdev-swap-2.28-7.el8_10.x86_64", "Installed: libblockdev-utils-2.28-7.el8_10.x86_64", "Installed: libudisks2-2.9.0-16.el8_10.1.x86_64", "Installed: libbytesize-1.4-3.el8.x86_64", "Installed: coreutils-8.30-17.el8_10.x86_64", "Installed: coreutils-common-8.30-17.el8_10.x86_64", "Installed: libcurl-7.61.1-34.el8_10.11.x86_64", "Installed: mdadm-4.2-19.el8_10.x86_64", "Installed: libnfsidmap-1:2.3.3-69.el8_10.x86_64", "Installed: curl-7.61.1-34.el8_10.11.x86_64", "Installed: volume_key-libs-0.3.11-6.el8.x86_64", "Installed: fwupd-1.7.8-2.el8.x86_64", "Installed: python3-libs-3.6.8-74.el8_10.x86_64", "Installed: gnutls-3.6.16-8.el8_10.5.x86_64", "Installed: libgcab1-1.1-1.el8.x86_64", "Installed: nfs-utils-1:2.3.3-69.el8_10.x86_64", "Installed: udisks2-2.9.0-16.el8_10.1.x86_64", "Installed: krb5-devel-1.18.2-33.el8_10.x86_64", "Installed: krb5-libs-1.18.2-33.el8_10.x86_64", "Installed: libgudev-232-4.el8.x86_64", "Installed: libgusb-0.3.0-1.el8.x86_64", "Installed: libxmlb-0.1.15-1.el8.x86_64", "Installed: dosfstools-4.1-6.el8.x86_64", "Removed: coreutils-8.30-16.el8_10.x86_64", "Removed: coreutils-common-8.30-16.el8_10.x86_64", "Removed: libcurl-7.61.1-34.el8_10.10.x86_64", "Removed: gnutls-3.6.16-8.el8_10.4.x86_64", "Removed: libnfsidmap-1:2.3.3-68.el8_10.x86_64", "Removed: platform-python-3.6.8-73.el8_10.x86_64", "Removed: curl-7.61.1-34.el8_10.10.x86_64", "Removed: krb5-devel-1.18.2-32.el8_10.x86_64", "Removed: krb5-libs-1.18.2-32.el8_10.x86_64", "Removed: dbxtool-8-5.el8_3.2.x86_64", "Removed: nfs-utils-1:2.3.3-68.el8_10.x86_64", "Removed: libkadm5-1.18.2-32.el8_10.x86_64", "Removed: python3-libs-3.6.8-73.el8_10.x86_64"], "results_file": "/root/.ansible_async/j303642143353.16025", "started": 1, "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [infra.leapp.common : update-and-reboot | Reboot when updates applied] **** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/update-and-reboot.yml:13 changed: [managed-node01] => {"changed": true, "elapsed": 120, "rebooted": true} TASK [leapp-upgrade | Create /etc/leapp/files/leapp_upgrade_repositories.repo] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tasks/leapp-upgrade.yml:25 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.upgrade : leapp-upgrade | Include disable-previous-repo-files.yml] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tasks/leapp-upgrade.yml:36 skipping: [managed-node01] => {"changed": false, "skipped_reason": "No items in the list"} TASK [leapp-upgrade | Unload kernel modules before upgrade] ******************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tasks/leapp-upgrade.yml:42 TASK [infra.leapp.common : manage_kernel_modules | Ensure ansible_facts used by role] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/manage_kernel_modules.yml:5 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/set_facts.yml for managed-node01 TASK [infra.leapp.common : set_facts | Ensure ansible_facts used by role] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/set_facts.yml:10 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : manage_kernel_modules | Load or unload kernel modules] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/manage_kernel_modules.yml:8 skipping: [managed-node01] => {"changed": false, "skipped_reason": "No items in the list"} TASK [infra.leapp.upgrade : leapp-upgrade | Start Leapp OS upgrade] ************ task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tasks/leapp-upgrade.yml:52 ASYNC FAILED on managed-node01: jid=j676870258004.1622 fatal: [managed-node01]: FAILED! => {"ansible_job_id": "j676870258004.1622", "changed": true, "cmd": "set -o pipefail; export PATH=$PATH; ulimit -n 16384; leapp upgrade --report-schema=1.2.0 2>&1 | tee -a /var/log/leapp/ansible_leapp_upgrade.log\n", "delta": "0:00:38.419874", "end": "2026-03-25 10:57:25.777740", "finished": 1, "msg": "non-zero return code", "rc": 1, "results_file": "/root/.ansible_async/j676870258004.1622", "start": "2026-03-25 10:56:47.357866", "started": 1, "stderr": "", "stderr_lines": [], "stdout": "==> Processing phase `configuration_phase`\n====> * ipu_workflow_config\n IPU workflow config actor\n==> Processing phase `FactsCollection`\n====> * scan_grub_config\n Scan grub configuration files for errors.\n====> * scan_target_os_image\n Scans the provided target OS ISO image to use as a content source for the IPU, if any.\n====> * nis_scanner\n Collect information about the NIS packages configuration.\n====> * scandasd\n In case of s390x architecture, check whether DASD is used.\n====> * repository_mapping\n Produces message containing repository mapping based on provided file.\n====> * transaction_workarounds\n Provides additional RPM transaction tasks based on bundled RPM packages.\n====> * system_facts\n Provides data about many facts from system.\n====> * udevadm_info\n Produces data exported by the \"udevadm info\" command.\n====> * rpm_scanner\n Provides data about installed RPM Packages.\n====> * scanzfcp\n In case of s390x architecture, check whether ZFCP is used.\n====> * scan_custom_modifications_actor\n Collects information about files in leapp directories that have been modified or newly added.\n====> * scan_source_files\n Scan files (explicitly specified) of the source system.\n====> * storage_scanner\n Provides data about storage settings.\n====> * ifcfg_scanner\n Scan ifcfg files with legacy network configuration\n====> * remove_obsolete_gpg_keys\n Remove obsoleted RPM GPG keys.\n====> * scan_pkg_manager\n Provides data about package manager (yum/dnf)\n====> * scan_custom_repofile\n Scan the custom /etc/leapp/files/leapp_upgrade_repositories.repo repo file.\n====> * firewalld_collect_used_object_names\n This actor reads firewalld's configuration and produces Model\n====> * roce_scanner\n Detect active RoCE NICs on IBM Z machines.\n====> * xorgdrvfacts8to9\n Check the journal logs for deprecated Xorg drivers.\n====> * register_ruby_irb_adjustment\n Register a workaround to allow rubygem-irb's directory -> symlink conversion.\n====> * scan_source_boot_entry\n Scan the default boot entry of the source system.\n====> * scan_subscription_manager_info\n Scans the current system for subscription manager information\n====> * get_installed_desktops\n Actor checks if kde or gnome desktop environments\n====> * scan_grub_device_name\n Find the name of the block devices where GRUB is located\n====> * get_enabled_modules\n Provides data about which module streams are enabled on the source system.\n====> * scan_systemd_source\n Provides info about systemd on the source system\n====> * copy_dnf_conf_into_target_userspace\n Copy dnf.conf into target userspace\n====> * root_scanner\n Scan the system root directory and produce a message containing\n====> * scan_files_for_target_userspace\n Scan the source system and identify files that will be copied into the target userspace when it is created.\n====> * open_ssl_config_scanner\n Read an OpenSSL configuration file for further analysis.\n====> * firewalld_collect_global_config\n This actor reads firewalld's configuration and produces Model\n====> * scanmemory\n Scan Memory of the machine.\n====> * read_openssh_config\n Collect information about the OpenSSH configuration.\n====> * scanclienablerepo\n Produce CustomTargetRepository based on the LEAPP_ENABLE_REPOS in config.\n====> * scancryptopolicies\n Scan information about system wide set crypto policies including:\n====> * sssd_facts_8to9\n Check SSSD configuration for changes in RHEL9 and report them in model.\n====> * scan_hybrid_image_azure\n Check if the system is using Azure hybrid image.\n====> * selinuxcontentscanner\n Scan the system for any SELinux customizations\n====> * network_manager_read_config\n Provides data about NetworkManager configuration.\n====> * scan_kernel_cmdline\n Scan the kernel command line of the booted system.\n====> * scanblacklistca\n Scan the file system for distrusted CA's in the blacklist directory.\n====> * checkrhui\n Check if system is using RHUI infrastructure (on public cloud) and send messages to\n====> * check_custom_network_scripts\n Check the existence of custom network-scripts and warn user about possible\n====> * network_manager_connection_scanner\n Scan NetworkManager connection keyfiles\n====> * luks_scanner\n Provides data about active LUKS devices.\n====> * persistentnetnames\n Get network interface information for physical ethernet interfaces of the original system.\n====> * load_device_driver_deprecation_data\n Loads deprecation data for drivers and devices (PCI & CPU)\n====> * scan_sap_hana\n Gathers information related to SAP HANA instances on the system.\n====> * scan_default_initramfs\n Scan details of the default boot entry's initramfs image.\n====> * scan_fips\n Determine whether the source system has FIPS enabled.\n====> * trusted_gpg_keys_scanner\n Scan for trusted GPG keys.\n====> * scancpu\n Scan CPUs of the machine.\n====> * distribution_signed_rpm_scanner\n Provide data about distribution signed & third-party RPM packages.\n====> * xfs_info_scanner\n This actor scans all mounted mountpoints for XFS information.\n====> * detect_kernel_drivers\n Matches all currently loaded kernel drivers against known deprecated and removed drivers.\n====> * used_repository_scanner\n Scan used enabled repositories\n====> * repositories_blacklist\n Exclude target repositories provided by Red Hat without support.\n====> * biosdevname\n Enable biosdevname on the target RHEL system if all interfaces on the source RHEL\n====> * rpm_transaction_config_tasks_collector\n Provides additional RPM transaction tasks from /etc/leapp/transaction.\n====> * persistentnetnamesdisable\n Disable systemd-udevd persistent network naming on machine with single eth0 NIC\n====> * satellite_upgrade_facts\n Report which Satellite packages require updates and how to handle PostgreSQL data\n====> * pci_devices_scanner\n Provides data about existing PCI Devices.\n====> * multipath_conf_read_8to9\n Read multipath configuration files and extract the necessary information\n====> * scan_source_kernel\n Scan the source system kernel.\n====> * ipa_scanner\n Scan system for ipa-client and ipa-server status\n====> * satellite_upgrade_services\n Reconfigure Satellite services\n====> * scan_dynamic_linker_configuration\n Scan the dynamic linker configuration and find modifications.\n====> * vdo_conversion_scanner\n Provides conversion info about VDO devices.\n====> * pes_events_scanner\n Provides data about package events from Package Evolution Service.\n====> * setuptargetrepos\n Produces list of repositories that should be available to be used during IPU process.\n\n============================================================\n ERRORS \n============================================================\n\n2026-03-25 10:57:07.145650 [ERROR] Actor: scan_subscription_manager_info\nMessage: A subscription-manager command failed to execute\nSummary:\n Details: Command ['subscription-manager', 'release'] failed with exit code 1.\n Stderr: This system is not yet registered. Try 'subscription-manager register --help' for more information.\n Hint: Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\n Link: https://access.redhat.com/solutions/6138372\n\n============================================================\n END OF ERRORS \n============================================================\n\nDebug output written to /var/log/leapp/leapp-upgrade.log\n\n============================================================\n REPORT OVERVIEW \n============================================================\n\nFollowing errors occurred and the upgrade cannot continue:\n 1. Actor: scan_subscription_manager_info\n Message: A subscription-manager command failed to execute\n\nHIGH and MEDIUM severity reports:\n 1. Packages available in excluded repositories will not be installed\n\nReports summary:\n Errors: 1\n Inhibitors: 0\n HIGH severity reports: 1\n MEDIUM severity reports: 0\n LOW severity reports: 0\n INFO severity reports: 1\n\nBefore continuing, review the full report below for details about discovered problems and possible remediation instructions:\n A report has been generated at /var/log/leapp/leapp-report.txt\n A report has been generated at /var/log/leapp/leapp-report.json\n\n============================================================\n END OF REPORT OVERVIEW \n============================================================\n\nAnswerfile has been generated at /var/log/leapp/answerfile", "stdout_lines": ["==> Processing phase `configuration_phase`", "====> * ipu_workflow_config", " IPU workflow config actor", "==> Processing phase `FactsCollection`", "====> * scan_grub_config", " Scan grub configuration files for errors.", "====> * scan_target_os_image", " Scans the provided target OS ISO image to use as a content source for the IPU, if any.", "====> * nis_scanner", " Collect information about the NIS packages configuration.", "====> * scandasd", " In case of s390x architecture, check whether DASD is used.", "====> * repository_mapping", " Produces message containing repository mapping based on provided file.", "====> * transaction_workarounds", " Provides additional RPM transaction tasks based on bundled RPM packages.", "====> * system_facts", " Provides data about many facts from system.", "====> * udevadm_info", " Produces data exported by the \"udevadm info\" command.", "====> * rpm_scanner", " Provides data about installed RPM Packages.", "====> * scanzfcp", " In case of s390x architecture, check whether ZFCP is used.", "====> * scan_custom_modifications_actor", " Collects information about files in leapp directories that have been modified or newly added.", "====> * scan_source_files", " Scan files (explicitly specified) of the source system.", "====> * storage_scanner", " Provides data about storage settings.", "====> * ifcfg_scanner", " Scan ifcfg files with legacy network configuration", "====> * remove_obsolete_gpg_keys", " Remove obsoleted RPM GPG keys.", "====> * scan_pkg_manager", " Provides data about package manager (yum/dnf)", "====> * scan_custom_repofile", " Scan the custom /etc/leapp/files/leapp_upgrade_repositories.repo repo file.", "====> * firewalld_collect_used_object_names", " This actor reads firewalld's configuration and produces Model", "====> * roce_scanner", " Detect active RoCE NICs on IBM Z machines.", "====> * xorgdrvfacts8to9", " Check the journal logs for deprecated Xorg drivers.", "====> * register_ruby_irb_adjustment", " Register a workaround to allow rubygem-irb's directory -> symlink conversion.", "====> * scan_source_boot_entry", " Scan the default boot entry of the source system.", "====> * scan_subscription_manager_info", " Scans the current system for subscription manager information", "====> * get_installed_desktops", " Actor checks if kde or gnome desktop environments", "====> * scan_grub_device_name", " Find the name of the block devices where GRUB is located", "====> * get_enabled_modules", " Provides data about which module streams are enabled on the source system.", "====> * scan_systemd_source", " Provides info about systemd on the source system", "====> * copy_dnf_conf_into_target_userspace", " Copy dnf.conf into target userspace", "====> * root_scanner", " Scan the system root directory and produce a message containing", "====> * scan_files_for_target_userspace", " Scan the source system and identify files that will be copied into the target userspace when it is created.", "====> * open_ssl_config_scanner", " Read an OpenSSL configuration file for further analysis.", "====> * firewalld_collect_global_config", " This actor reads firewalld's configuration and produces Model", "====> * scanmemory", " Scan Memory of the machine.", "====> * read_openssh_config", " Collect information about the OpenSSH configuration.", "====> * scanclienablerepo", " Produce CustomTargetRepository based on the LEAPP_ENABLE_REPOS in config.", "====> * scancryptopolicies", " Scan information about system wide set crypto policies including:", "====> * sssd_facts_8to9", " Check SSSD configuration for changes in RHEL9 and report them in model.", "====> * scan_hybrid_image_azure", " Check if the system is using Azure hybrid image.", "====> * selinuxcontentscanner", " Scan the system for any SELinux customizations", "====> * network_manager_read_config", " Provides data about NetworkManager configuration.", "====> * scan_kernel_cmdline", " Scan the kernel command line of the booted system.", "====> * scanblacklistca", " Scan the file system for distrusted CA's in the blacklist directory.", "====> * checkrhui", " Check if system is using RHUI infrastructure (on public cloud) and send messages to", "====> * check_custom_network_scripts", " Check the existence of custom network-scripts and warn user about possible", "====> * network_manager_connection_scanner", " Scan NetworkManager connection keyfiles", "====> * luks_scanner", " Provides data about active LUKS devices.", "====> * persistentnetnames", " Get network interface information for physical ethernet interfaces of the original system.", "====> * load_device_driver_deprecation_data", " Loads deprecation data for drivers and devices (PCI & CPU)", "====> * scan_sap_hana", " Gathers information related to SAP HANA instances on the system.", "====> * scan_default_initramfs", " Scan details of the default boot entry's initramfs image.", "====> * scan_fips", " Determine whether the source system has FIPS enabled.", "====> * trusted_gpg_keys_scanner", " Scan for trusted GPG keys.", "====> * scancpu", " Scan CPUs of the machine.", "====> * distribution_signed_rpm_scanner", " Provide data about distribution signed & third-party RPM packages.", "====> * xfs_info_scanner", " This actor scans all mounted mountpoints for XFS information.", "====> * detect_kernel_drivers", " Matches all currently loaded kernel drivers against known deprecated and removed drivers.", "====> * used_repository_scanner", " Scan used enabled repositories", "====> * repositories_blacklist", " Exclude target repositories provided by Red Hat without support.", "====> * biosdevname", " Enable biosdevname on the target RHEL system if all interfaces on the source RHEL", "====> * rpm_transaction_config_tasks_collector", " Provides additional RPM transaction tasks from /etc/leapp/transaction.", "====> * persistentnetnamesdisable", " Disable systemd-udevd persistent network naming on machine with single eth0 NIC", "====> * satellite_upgrade_facts", " Report which Satellite packages require updates and how to handle PostgreSQL data", "====> * pci_devices_scanner", " Provides data about existing PCI Devices.", "====> * multipath_conf_read_8to9", " Read multipath configuration files and extract the necessary information", "====> * scan_source_kernel", " Scan the source system kernel.", "====> * ipa_scanner", " Scan system for ipa-client and ipa-server status", "====> * satellite_upgrade_services", " Reconfigure Satellite services", "====> * scan_dynamic_linker_configuration", " Scan the dynamic linker configuration and find modifications.", "====> * vdo_conversion_scanner", " Provides conversion info about VDO devices.", "====> * pes_events_scanner", " Provides data about package events from Package Evolution Service.", "====> * setuptargetrepos", " Produces list of repositories that should be available to be used during IPU process.", "", "============================================================", " ERRORS ", "============================================================", "", "2026-03-25 10:57:07.145650 [ERROR] Actor: scan_subscription_manager_info", "Message: A subscription-manager command failed to execute", "Summary:", " Details: Command ['subscription-manager', 'release'] failed with exit code 1.", " Stderr: This system is not yet registered. Try 'subscription-manager register --help' for more information.", " Hint: Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.", " Link: https://access.redhat.com/solutions/6138372", "", "============================================================", " END OF ERRORS ", "============================================================", "", "Debug output written to /var/log/leapp/leapp-upgrade.log", "", "============================================================", " REPORT OVERVIEW ", "============================================================", "", "Following errors occurred and the upgrade cannot continue:", " 1. Actor: scan_subscription_manager_info", " Message: A subscription-manager command failed to execute", "", "HIGH and MEDIUM severity reports:", " 1. Packages available in excluded repositories will not be installed", "", "Reports summary:", " Errors: 1", " Inhibitors: 0", " HIGH severity reports: 1", " MEDIUM severity reports: 0", " LOW severity reports: 0", " INFO severity reports: 1", "", "Before continuing, review the full report below for details about discovered problems and possible remediation instructions:", " A report has been generated at /var/log/leapp/leapp-report.txt", " A report has been generated at /var/log/leapp/leapp-report.json", "", "============================================================", " END OF REPORT OVERVIEW ", "============================================================", "", "Answerfile has been generated at /var/log/leapp/answerfile"]} TASK [leapp-upgrade | Run parse_leapp_report to check for inhibitors] ********** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tasks/leapp-upgrade.yml:68 TASK [infra.leapp.common : parse_leapp_report | Ensure ansible_facts used by role] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:4 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/set_facts.yml for managed-node01 TASK [infra.leapp.common : set_facts | Ensure ansible_facts used by role] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/set_facts.yml:10 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : parse_leapp_report | Default upgrade_inhibited to false] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:15 ok: [managed-node01] => {"ansible_facts": {"upgrade_inhibited": false}, "changed": false} TASK [infra.leapp.common : parse_leapp_report | Fail if pre-upgrade analysis was not run] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:20 ok: [managed-node01] => {"changed": false, "failed_when_result": false, "stat": {"atime": 1774450644.8255649, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "554a6a04b635bd00600ae28d31f0847128b10c8e", "ctime": 1774450644.8255649, "dev": 51715, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 452984986, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1774450644.8255649, "nlink": 1, "path": "/var/log/leapp/leapp-report.txt", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 2702, "uid": 0, "version": "3652298052", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false}} TASK [infra.leapp.common : parse_leapp_report | Collect human readable report results] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:26 ok: [managed-node01] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false} TASK [infra.leapp.common : parse_leapp_report | Collect JSON report results] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:32 ok: [managed-node01] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false} TASK [infra.leapp.common : parse_leapp_report | Parse report results] ********** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:38 ok: [managed-node01] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false} TASK [infra.leapp.common : parse_leapp_report | Clear leapp_inhibitors] ******** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:45 ok: [managed-node01] => {"ansible_facts": {"leapp_inhibitors": []}, "changed": false} TASK [infra.leapp.common : parse_leapp_report | Check for inhibitors] ********** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:49 ok: [managed-node01] => (item={'audience': 'sysadmin', 'groups': ['error'], 'key': '7ec8269784db1bba2ac54ae438689ef397e16833', 'severity': 'high', 'summary': '{"details": "Command [\'subscription-manager\', \'release\'] failed with exit code 1.", "stderr": "This system is not yet registered. Try \'subscription-manager register --help\' for more information.\\n", "hint": "Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.", "link": "https://access.redhat.com/solutions/6138372"}', 'title': 'A subscription-manager command failed to execute', 'timeStamp': '2026-03-25T14:57:07.145864Z', 'hostname': 'managed-node01', 'actor': 'scan_subscription_manager_info', 'id': '305c341f4b0a5612edfe5c9cd2135d82287adf5adc8c15c3f34be5f5f44fbb82'}) => {"ansible_facts": {"leapp_inhibitors": [{"actor": "scan_subscription_manager_info", "audience": "sysadmin", "groups": ["error"], "hostname": "managed-node01", "id": "305c341f4b0a5612edfe5c9cd2135d82287adf5adc8c15c3f34be5f5f44fbb82", "key": "7ec8269784db1bba2ac54ae438689ef397e16833", "severity": "high", "summary": "{\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "timeStamp": "2026-03-25T14:57:07.145864Z", "title": "A subscription-manager command failed to execute"}], "upgrade_inhibited": true}, "ansible_loop_var": "item", "changed": false, "item": {"actor": "scan_subscription_manager_info", "audience": "sysadmin", "groups": ["error"], "hostname": "managed-node01", "id": "305c341f4b0a5612edfe5c9cd2135d82287adf5adc8c15c3f34be5f5f44fbb82", "key": "7ec8269784db1bba2ac54ae438689ef397e16833", "severity": "high", "summary": "{\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "timeStamp": "2026-03-25T14:57:07.145864Z", "title": "A subscription-manager command failed to execute"}} skipping: [managed-node01] => (item={'audience': 'sysadmin', 'detail': {'remediations': [{'context': 'If some of excluded repositories are still required to be used during the upgrade, execute leapp with the --enablerepo option with the repoid of the repository required to be enabled as an argument (the option can be used multiple times).', 'type': 'hint'}]}, 'groups': ['repository', 'failure'], 'key': '1b9132cb2362ae7830e48eee7811be9527747de8', 'severity': 'info', 'summary': 'The following repositories are not supported by Red Hat and are excluded from the list of repositories used during the upgrade.\n- codeready-builder-beta-for-rhel-9-x86_64-rpms\n- codeready-builder-beta-for-rhel-9-aarch64-rpms\n- rhui-codeready-builder-for-rhel-9-aarch64-rhui-rpms\n- codeready-builder-for-rhel-9-s390x-rpms\n- codeready-builder-for-rhel-9-s390x-eus-rpms\n- codeready-builder-for-rhel-9-rhui-rpms\n- codeready-builder-for-rhel-9-aarch64-rpms\n- codeready-builder-for-rhel-9-ppc64le-eus-rpms\n- codeready-builder-for-rhel-9-aarch64-eus-rpms\n- codeready-builder-beta-for-rhel-9-ppc64le-rpms\n- codeready-builder-for-rhel-9-ppc64le-rpms\n- crb\n- codeready-builder-beta-for-rhel-9-s390x-rpms\n- codeready-builder-for-rhel-9-x86_64-rpms\n- codeready-builder-for-rhel-9-x86_64-rhui-rpms\n- rhui-codeready-builder-for-rhel-9-x86_64-rhui-rpms\n- codeready-builder-for-rhel-9-x86_64-eus-rpms', 'title': 'Excluded target system repositories', 'timeStamp': '2026-03-25T14:57:21.239058Z', 'hostname': 'managed-node01', 'actor': 'repositories_blacklist', 'id': '3f336ed20b6442af36d5d6989493dc9e96b9fff47b2725c66cdcd32c8f6b29f6'}) => {"ansible_loop_var": "item", "changed": false, "item": {"actor": "repositories_blacklist", "audience": "sysadmin", "detail": {"remediations": [{"context": "If some of excluded repositories are still required to be used during the upgrade, execute leapp with the --enablerepo option with the repoid of the repository required to be enabled as an argument (the option can be used multiple times).", "type": "hint"}]}, "groups": ["repository", "failure"], "hostname": "managed-node01", "id": "3f336ed20b6442af36d5d6989493dc9e96b9fff47b2725c66cdcd32c8f6b29f6", "key": "1b9132cb2362ae7830e48eee7811be9527747de8", "severity": "info", "summary": "The following repositories are not supported by Red Hat and are excluded from the list of repositories used during the upgrade.\n- codeready-builder-beta-for-rhel-9-x86_64-rpms\n- codeready-builder-beta-for-rhel-9-aarch64-rpms\n- rhui-codeready-builder-for-rhel-9-aarch64-rhui-rpms\n- codeready-builder-for-rhel-9-s390x-rpms\n- codeready-builder-for-rhel-9-s390x-eus-rpms\n- codeready-builder-for-rhel-9-rhui-rpms\n- codeready-builder-for-rhel-9-aarch64-rpms\n- codeready-builder-for-rhel-9-ppc64le-eus-rpms\n- codeready-builder-for-rhel-9-aarch64-eus-rpms\n- codeready-builder-beta-for-rhel-9-ppc64le-rpms\n- codeready-builder-for-rhel-9-ppc64le-rpms\n- crb\n- codeready-builder-beta-for-rhel-9-s390x-rpms\n- codeready-builder-for-rhel-9-x86_64-rpms\n- codeready-builder-for-rhel-9-x86_64-rhui-rpms\n- rhui-codeready-builder-for-rhel-9-x86_64-rhui-rpms\n- codeready-builder-for-rhel-9-x86_64-eus-rpms", "timeStamp": "2026-03-25T14:57:21.239058Z", "title": "Excluded target system repositories"}, "skip_reason": "Conditional result was False"} skipping: [managed-node01] => (item={'audience': 'sysadmin', 'detail': {'related_resources': [{'scheme': 'package', 'title': 'jitterentropy-devel'}]}, 'groups': ['repository'], 'key': '2437e204808f987477c0e9be8e4c95b3a87a9f3e', 'severity': 'high', 'summary': '1 packages will be skipped because they are available only in target system repositories that are intentionally excluded from the list of repositories used during the upgrade. See the report message titled "Excluded target system repositories" for details.\nThe list of these packages:\n- jitterentropy-devel (repoid: codeready-builder-for-rhel-9-x86_64-rpms)', 'title': 'Packages available in excluded repositories will not be installed', 'timeStamp': '2026-03-25T14:57:24.660315Z', 'hostname': 'managed-node01', 'actor': 'pes_events_scanner', 'id': 'e02ff8695ba8e020789a9528e82b29e8353a437d078eed082b4b7dfea0858ec7'}) => {"ansible_loop_var": "item", "changed": false, "item": {"actor": "pes_events_scanner", "audience": "sysadmin", "detail": {"related_resources": [{"scheme": "package", "title": "jitterentropy-devel"}]}, "groups": ["repository"], "hostname": "managed-node01", "id": "e02ff8695ba8e020789a9528e82b29e8353a437d078eed082b4b7dfea0858ec7", "key": "2437e204808f987477c0e9be8e4c95b3a87a9f3e", "severity": "high", "summary": "1 packages will be skipped because they are available only in target system repositories that are intentionally excluded from the list of repositories used during the upgrade. See the report message titled \"Excluded target system repositories\" for details.\nThe list of these packages:\n- jitterentropy-devel (repoid: codeready-builder-for-rhel-9-x86_64-rpms)", "timeStamp": "2026-03-25T14:57:24.660315Z", "title": "Packages available in excluded repositories will not be installed"}, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : parse_leapp_report | Collect inhibitors] ************ task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:61 ok: [managed-node01] => {"changed": false, "cmd": ["awk", "/\\(inhibitor\\)/,/^-------/", "/var/log/leapp/leapp-report.txt"], "delta": "0:00:00.003824", "end": "2026-03-25 10:57:49.756416", "failed_when_result": false, "msg": "", "rc": 0, "start": "2026-03-25 10:57:49.752592", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [infra.leapp.common : parse_leapp_report | Collect high errors] *********** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:70 ok: [managed-node01] => {"changed": false, "cmd": ["awk", "/high \\(error\\)/,/^-------/", "/var/log/leapp/leapp-report.txt"], "delta": "0:00:00.003736", "end": "2026-03-25 10:57:50.102300", "failed_when_result": false, "msg": "", "rc": 0, "start": "2026-03-25 10:57:50.098564", "stderr": "", "stderr_lines": [], "stdout": "Risk Factor: high (error)\nTitle: A subscription-manager command failed to execute\nSummary: {\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}\nKey: 7ec8269784db1bba2ac54ae438689ef397e16833\n----------------------------------------", "stdout_lines": ["Risk Factor: high (error)", "Title: A subscription-manager command failed to execute", "Summary: {\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "Key: 7ec8269784db1bba2ac54ae438689ef397e16833", "----------------------------------------"]} TASK [infra.leapp.upgrade : leapp-upgrade | Display inhibitors] **************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tasks/leapp-upgrade.yml:73 skipping: [managed-node01] => {} TASK [infra.leapp.upgrade : leapp-upgrade | Display errors] ******************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tasks/leapp-upgrade.yml:78 ok: [managed-node01] => { "results_errors.stdout_lines": [ "Risk Factor: high (error)", "Title: A subscription-manager command failed to execute", "Summary: {\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "Key: 7ec8269784db1bba2ac54ae438689ef397e16833", "----------------------------------------" ] } TASK [infra.leapp.upgrade : leapp-upgrade | Fail Leapp upgrade] **************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tasks/leapp-upgrade.yml:83 fatal: [managed-node01]: FAILED! => {"changed": false, "msg": "Errors encountered running Leapp upgrade command. Review the tasks above or the result file at /var/log/leapp/leapp-report.txt."} TASK [Copy reports to the controller] ****************************************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tasks/main.yml:40 TASK [infra.leapp.common : copy_reports_to_controller | Ensure ansible_facts used by role] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_reports_to_controller.yml:20 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/set_facts.yml for managed-node01 TASK [infra.leapp.common : set_facts | Ensure ansible_facts used by role] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/set_facts.yml:10 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : copy_reports_to_controller | Ensure reports directory on controller] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_reports_to_controller.yml:23 changed: [managed-node01 -> localhost] => {"changed": true, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tests/ansible_leapp_upgrade_logs_2026-03-25_14-53-32", "secontext": "unconfined_u:object_r:admin_home_t:s0", "size": 6, "state": "directory", "uid": 0} TASK [infra.leapp.common : copy_reports_to_controller | Fetch report files if they exist] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_reports_to_controller.yml:33 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/fetch_file_if_exists.yml for managed-node01 => (item=/var/log/leapp/leapp-report.txt) included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/fetch_file_if_exists.yml for managed-node01 => (item=/var/log/leapp/leapp-report.json) included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/fetch_file_if_exists.yml for managed-node01 => (item=/var/log/leapp/leapp-upgrade.log) TASK [infra.leapp.common : fetch_file_if_exists | Check if file exists] ******** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/fetch_file_if_exists.yml:7 ok: [managed-node01] => {"changed": false, "stat": {"atime": 1774450668.4387376, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "554a6a04b635bd00600ae28d31f0847128b10c8e", "ctime": 1774450644.8255649, "dev": 51715, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 452984986, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1774450644.8255649, "nlink": 1, "path": "/var/log/leapp/leapp-report.txt", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 2702, "uid": 0, "version": "3652298052", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false}} TASK [infra.leapp.common : fetch_file_if_exists | Copy report file to the controller] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/fetch_file_if_exists.yml:12 changed: [managed-node01] => {"changed": true, "checksum": "554a6a04b635bd00600ae28d31f0847128b10c8e", "dest": "/root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tests/ansible_leapp_upgrade_logs_2026-03-25_14-53-32/managed-node01/leapp-report.txt", "md5sum": "b78f083fff320e5ef1a64de9c61dce3c", "remote_checksum": "554a6a04b635bd00600ae28d31f0847128b10c8e", "remote_md5sum": null} TASK [infra.leapp.common : fetch_file_if_exists | Check if file exists] ******** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/fetch_file_if_exists.yml:7 ok: [managed-node01] => {"changed": false, "stat": {"atime": 1774450669.2637439, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "3fddc697e29b0315d16ade1af2b09a94f5431826", "ctime": 1774450644.8255649, "dev": 51715, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 452984987, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1774450644.8255649, "nlink": 1, "path": "/var/log/leapp/leapp-report.json", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 4007, "uid": 0, "version": "254706288", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false}} TASK [infra.leapp.common : fetch_file_if_exists | Copy report file to the controller] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/fetch_file_if_exists.yml:12 changed: [managed-node01] => {"changed": true, "checksum": "3fddc697e29b0315d16ade1af2b09a94f5431826", "dest": "/root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tests/ansible_leapp_upgrade_logs_2026-03-25_14-53-32/managed-node01/leapp-report.json", "md5sum": "42d6f41ba63ef5720f28fd361de9fc5c", "remote_checksum": "3fddc697e29b0315d16ade1af2b09a94f5431826", "remote_md5sum": null} TASK [infra.leapp.common : fetch_file_if_exists | Check if file exists] ******** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/fetch_file_if_exists.yml:7 ok: [managed-node01] => {"changed": false, "stat": {"atime": 1774450609.0862637, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 3160, "charset": "us-ascii", "checksum": "df6095f85b38f85104e6e9eac4a6af8a3c9b2ef7", "ctime": 1774450644.8265648, "dev": 51715, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 452984983, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1774450644.8265648, "nlink": 1, "path": "/var/log/leapp/leapp-upgrade.log", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1616909, "uid": 0, "version": "1604346828", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false}} TASK [infra.leapp.common : fetch_file_if_exists | Copy report file to the controller] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/fetch_file_if_exists.yml:12 changed: [managed-node01] => {"changed": true, "checksum": "df6095f85b38f85104e6e9eac4a6af8a3c9b2ef7", "dest": "/root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tests/ansible_leapp_upgrade_logs_2026-03-25_14-53-32/managed-node01/leapp-upgrade.log", "md5sum": "f941ad4fab8ecf188ae5d19b02945f36", "remote_checksum": "df6095f85b38f85104e6e9eac4a6af8a3c9b2ef7", "remote_md5sum": null} TASK [copy_reports_to_controller | Copy log file to the controller] ************ task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_reports_to_controller.yml:42 TASK [infra.leapp.common : copy_archive_leapp_log | Ensure ansible_facts used by role] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_archive_leapp_log.yml:11 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/set_facts.yml for managed-node01 TASK [infra.leapp.common : set_facts | Ensure ansible_facts used by role] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/set_facts.yml:10 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : copy_archive_leapp_log | Check for log file] ******** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_archive_leapp_log.yml:14 ok: [managed-node01] => {"changed": false, "stat": {"atime": 1774450418.2718189, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 24, "charset": "us-ascii", "checksum": "62c67f99706c8625014b938b173b1c8914e4aed1", "ctime": 1774450645.6935713, "dev": 51715, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 562036890, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1774450645.6935713, "nlink": 1, "path": "/var/log/leapp/ansible_leapp_upgrade.log", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 9389, "uid": 0, "version": "3518586328", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false}} TASK [infra.leapp.common : copy_archive_leapp_log | Add end time to log file] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_archive_leapp_log.yml:22 changed: [managed-node01] => {"backup": "", "changed": true, "msg": "line added"} TASK [infra.leapp.common : copy_archive_leapp_log | Slurp file /var/log/leapp/ansible_leapp_upgrade.log] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_archive_leapp_log.yml:30 ok: [managed-node01] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false} TASK [infra.leapp.common : copy_archive_leapp_log | Decode file /var/log/leapp/ansible_leapp_upgrade.log] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_archive_leapp_log.yml:36 ok: [managed-node01] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false} TASK [infra.leapp.common : copy_archive_leapp_log | Ensure reports directory on controller] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_archive_leapp_log.yml:41 ok: [managed-node01 -> localhost] => {"changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tests/ansible_leapp_upgrade_logs_2026-03-25_14-53-32", "secontext": "unconfined_u:object_r:admin_home_t:s0", "size": 28, "state": "directory", "uid": 0} TASK [infra.leapp.common : copy_archive_leapp_log | Copy ansible leapp log to the controller] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_archive_leapp_log.yml:51 changed: [managed-node01] => {"changed": true, "checksum": "b6ab4d60d22ad15c0c5f92b9d71b0672f4df4a5a", "dest": "/root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tests/ansible_leapp_upgrade_logs_2026-03-25_14-53-32/managed-node01/ansible_leapp_upgrade.log", "md5sum": "e16fbc65b10e521d148539eb5b818779", "remote_checksum": "b6ab4d60d22ad15c0c5f92b9d71b0672f4df4a5a", "remote_md5sum": null} TASK [infra.leapp.common : copy_archive_leapp_log | Copy log file to timestamped location] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_archive_leapp_log.yml:57 changed: [managed-node01] => {"changed": true, "checksum": "b6ab4d60d22ad15c0c5f92b9d71b0672f4df4a5a", "dest": "/var/log/leapp/ansible_leapp_upgrade_2026-03-25_14-53-32.log", "gid": 0, "group": "root", "md5sum": "e16fbc65b10e521d148539eb5b818779", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:var_log_t:s0", "size": 9423, "src": "/var/log/leapp/ansible_leapp_upgrade.log", "state": "file", "uid": 0} TASK [infra.leapp.common : copy_archive_leapp_log | Remove original log file] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_archive_leapp_log.yml:69 changed: [managed-node01] => {"changed": true, "path": "/var/log/leapp/ansible_leapp_upgrade.log", "state": "absent"} TASK [Test | Check error] ****************************************************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tests/tests_default.yml:14 ok: [managed-node01] => { "msg": "errors {\n \"_ansible_no_log\": false,\n \"changed\": false,\n \"failed\": true,\n \"msg\": \"Errors encountered running Leapp upgrade command. Review the tasks above or the result file at /var/log/leapp/leapp-report.txt.\"\n}" } TASK [Test | Ensure correct error] ********************************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tests/tests_default.yml:18 ok: [managed-node01] => { "changed": false, "msg": "All assertions passed" } TASK [Test | Include cleanup logs] ********************************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/upgrade/tests/tests_default.yml:27 TASK [infra.leapp.common : cleanup_logs | Ensure ansible_facts used by role] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/cleanup_logs.yml:2 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/set_facts.yml for managed-node01 TASK [infra.leapp.common : set_facts | Ensure ansible_facts used by role] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/set_facts.yml:10 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : cleanup_logs | Cleanup | Remove log files] ********** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/cleanup_logs.yml:5 changed: [managed-node01] => {"changed": true, "cmd": "set -euxo pipefail\nrm -f /var/log/leapp/*.log\nrm -f /var/log/leapp/*.json\nrm -f /var/log/leapp/*.txt\n", "delta": "0:00:00.008123", "end": "2026-03-25 10:57:56.317553", "msg": "", "rc": 0, "start": "2026-03-25 10:57:56.309430", "stderr": "+ rm -f /var/log/leapp/ansible_leapp_upgrade_2026-03-25_14-53-32.log /var/log/leapp/leapp-upgrade.log\n+ rm -f /var/log/leapp/leapp-report.json\n+ rm -f /var/log/leapp/leapp-report.txt", "stderr_lines": ["+ rm -f /var/log/leapp/ansible_leapp_upgrade_2026-03-25_14-53-32.log /var/log/leapp/leapp-upgrade.log", "+ rm -f /var/log/leapp/leapp-report.json", "+ rm -f /var/log/leapp/leapp-report.txt"], "stdout": "", "stdout_lines": []} PLAY RECAP ********************************************************************* managed-node01 : ok=55 changed=15 unreachable=0 failed=0 skipped=19 rescued=2 ignored=0