[WARNING]: Collection infra.leapp does not support Ansible version 2.14.18 [WARNING]: running playbook inside collection infra.leapp ansible-playbook [core 2.14.18] config file = /etc/ansible/ansible.cfg configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.9/site-packages/ansible ansible collection location = /root/.ansible/collections:/usr/share/ansible/collections executable location = /usr/bin/ansible-playbook python version = 3.9.25 (main, Jan 14 2026, 00:00:00) [GCC 11.5.0 20240719 (Red Hat 11.5.0-11)] (/usr/bin/python3) jinja version = 3.1.2 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_default.yml **************************************************** 1 plays in /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tests/tests_default.yml PLAY [Test] ******************************************************************** TASK [Gathering Facts] ********************************************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tests/tests_default.yml:2 ok: [managed-node01] TASK [Test | Run role analysis] ************************************************ task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tests/tests_default.yml:9 TASK [infra.leapp.common : Log directory exists] ******************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:3 changed: [managed-node01] => {"changed": true, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/var/log/ripu", "secontext": "unconfined_u:object_r:var_log_t:s0", "size": 6, "state": "directory", "uid": 0} TASK [infra.leapp.common : Check for existing log file] ************************ task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:11 ok: [managed-node01] => {"changed": false, "stat": {"exists": false}} TASK [infra.leapp.common : Fail if log file already exists] ******************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:16 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : Create new log file] ******************************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:23 NOTIFIED HANDLER infra.leapp.common : Check for log file for managed-node01 NOTIFIED HANDLER infra.leapp.common : Add end time to log file for managed-node01 NOTIFIED HANDLER infra.leapp.common : Slurp ripu.log file for managed-node01 NOTIFIED HANDLER infra.leapp.common : Decode ripu.log file for managed-node01 NOTIFIED HANDLER infra.leapp.common : Rename log file for managed-node01 changed: [managed-node01] => {"changed": true, "checksum": "59136dfcef86eeb56e4cbe8c783de491c0e0f501", "dest": "/var/log/ripu/ripu.log", "gid": 0, "group": "root", "md5sum": "d4ad8d4bce6f271c6c2a2ae00562ff32", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:var_log_t:s0", "size": 61, "src": "/root/.ansible/tmp/ansible-tmp-1771331207.3767955-5693-71201330276967/source", "state": "file", "uid": 0} TASK [infra.leapp.common : /etc/ansible/facts.d directory exists] ************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:35 changed: [managed-node01] => {"changed": true, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/ansible/facts.d", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0} TASK [infra.leapp.common : Capture current ansible_facts for validation after upgrade] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:43 changed: [managed-node01] => {"changed": true, "checksum": "63939e8b278ceb3bc0722a169cbdd27ad14d45cc", "dest": "/etc/ansible/facts.d/pre_ripu.fact", "gid": 0, "group": "root", "md5sum": "1a90504643b2a59f9cc3ddd755895d93", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 13831, "src": "/root/.ansible/tmp/ansible-tmp-1771331208.4693317-5721-17461701937088/source", "state": "file", "uid": 0} TASK [infra.leapp.common : Capture a list of non-rhel versioned packages] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:51 ok: [managed-node01] => {"changed": false, "cmd": "set -o pipefail; export PATH=$PATH; rpm -qa | grep -ve '[\\.|+]el8' | grep -vE '^(gpg-pubkey|libmodulemd|katello-ca-consumer)' | sort", "delta": "0:00:00.848600", "end": "2026-02-17 07:26:50.288316", "failed_when_result": false, "msg": "non-zero return code", "rc": 1, "start": "2026-02-17 07:26:49.439716", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [infra.leapp.common : Create fact with the non-rhel versioned packages list] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:65 ok: [managed-node01] => {"ansible_facts": {"non_rhel_packages": []}, "changed": false} TASK [infra.leapp.common : Capture the list of non-rhel versioned packages in a separate fact file] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:69 changed: [managed-node01] => {"changed": true, "checksum": "97d170e1550eee4afc0af065b78cda302a97674c", "dest": "/etc/ansible/facts.d/non_rhel_packages.fact", "gid": 0, "group": "root", "md5sum": "d751713988987e9331980363e24189ce", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 2, "src": "/root/.ansible/tmp/ansible-tmp-1771331210.3934922-5754-242075813388270/source", "state": "file", "uid": 0} TASK [infra.leapp.analysis : Include tasks for preupg assistant analysis] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/main.yml:9 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : Include tasks for leapp preupgrade analysis] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/main.yml:13 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml for managed-node01 TASK [analysis-leapp | Register with Satellite activation key] ***************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:2 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [analysis-leapp | Include custom_local_repos for local_repos_pre_leapp] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:11 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : analysis-leapp | Install packages for preupgrade analysis on RHEL 7] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:19 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : analysis-leapp | Install packages for preupgrade analysis on RHEL 8] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:26 changed: [managed-node01] => {"changed": true, "msg": "", "rc": 0, "results": ["Installed: policycoreutils-python-utils-2.9-26.el8_10.noarch", "Installed: leapp-0.20.0-1.el8_10.noarch", "Installed: leapp-deps-0.20.0-1.el8_10.noarch", "Installed: leapp-upgrade-el8toel9-0.23.0-1.el8_10.noarch", "Installed: leapp-upgrade-el8toel9-deps-0.23.0-1.el8_10.noarch", "Installed: systemd-container-239-82.el8_10.15.x86_64", "Installed: python3-leapp-0.20.0-1.el8_10.noarch"]} TASK [infra.leapp.analysis : analysis-leapp | Install packages for preupgrade analysis on RHEL 9] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:33 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : analysis-leapp | Ensure leapp log directory exists] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:40 ok: [managed-node01] => {"changed": false, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/var/log/leapp", "secontext": "system_u:object_r:var_log_t:s0", "size": 6, "state": "directory", "uid": 0} TASK [infra.leapp.analysis : analysis-leapp | Populate leapp_answers file] ***** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:48 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [analysis-leapp | Create /etc/leapp/files/leapp_upgrade_repositories.repo] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:57 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : analysis-leapp | Leapp preupgrade report] ********* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:68 ASYNC FAILED on managed-node01: jid=j3410284144.6621 changed: [managed-node01] => {"ansible_job_id": "j3410284144.6621", "changed": true, "cmd": "set -o pipefail; export PATH=$PATH; ulimit -n 16384; leapp preupgrade --report-schema=1.2.0 2>&1 | tee -a /var/log/ripu/ripu.log\n", "delta": "0:00:39.086220", "end": "2026-02-17 07:27:39.709292", "failed_when_result": false, "finished": 1, "msg": "non-zero return code", "rc": 1, "results_file": "/root/.ansible_async/j3410284144.6621", "start": "2026-02-17 07:27:00.623072", "started": 1, "stderr": "", "stderr_lines": [], "stdout": "==> Processing phase `configuration_phase`\n====> * ipu_workflow_config\n IPU workflow config actor\n==> Processing phase `FactsCollection`\n====> * scan_target_os_image\n Scans the provided target OS ISO image to use as a content source for the IPU, if any.\n====> * scan_grub_config\n Scan grub configuration files for errors.\n====> * scancryptopolicies\n Scan information about system wide set crypto policies including:\n====> * udevadm_info\n Produces data exported by the \"udevadm info\" command.\n====> * transaction_workarounds\n Provides additional RPM transaction tasks based on bundled RPM packages.\n====> * repository_mapping\n Produces message containing repository mapping based on provided file.\n====> * scan_custom_modifications_actor\n Collects information about files in leapp directories that have been modified or newly added.\n====> * scandasd\n In case of s390x architecture, check whether DASD is used.\n====> * storage_scanner\n Provides data about storage settings.\n====> * network_manager_read_config\n Provides data about NetworkManager configuration.\n====> * ifcfg_scanner\n Scan ifcfg files with legacy network configuration\n====> * network_manager_connection_scanner\n Scan NetworkManager connection keyfiles\n====> * scan_files_for_target_userspace\n Scan the source system and identify files that will be copied into the target userspace when it is created.\n====> * scan_kernel_cmdline\n Scan the kernel command line of the booted system.\n====> * copy_dnf_conf_into_target_userspace\n Copy dnf.conf into target userspace\n====> * scanclienablerepo\n Produce CustomTargetRepository based on the LEAPP_ENABLE_REPOS in config.\n====> * firewalld_collect_used_object_names\n This actor reads firewalld's configuration and produces Model\n====> * system_facts\n Provides data about many facts from system.\n====> * scanblacklistca\n Scan the file system for distrusted CA's in the blacklist directory.\n====> * scan_pkg_manager\n Provides data about package manager (yum/dnf)\n====> * scanzfcp\n In case of s390x architecture, check whether ZFCP is used.\n====> * scan_source_files\n Scan files (explicitly specified) of the source system.\n====> * luks_scanner\n Provides data about active LUKS devices.\n====> * get_enabled_modules\n Provides data about which module streams are enabled on the source system.\n====> * scan_subscription_manager_info\n Scans the current system for subscription manager information\n====> * scan_source_boot_entry\n Scan the default boot entry of the source system.\n====> * firewalld_collect_global_config\n This actor reads firewalld's configuration and produces Model\n====> * scan_fips\n Determine whether the source system has FIPS enabled.\n====> * scan_grub_device_name\n Find the name of the block devices where GRUB is located\n====> * nis_scanner\n Collect information about the NIS packages configuration.\n====> * xorgdrvfacts8to9\n Check the journal logs for deprecated Xorg drivers.\n====> * scan_systemd_source\n Provides info about systemd on the source system\n====> * register_ruby_irb_adjustment\n Register a workaround to allow rubygem-irb's directory -> symlink conversion.\n====> * sssd_facts_8to9\n Check SSSD configuration for changes in RHEL9 and report them in model.\n====> * open_ssl_config_scanner\n Read an OpenSSL configuration file for further analysis.\n====> * scanmemory\n Scan Memory of the machine.\n====> * rpm_scanner\n Provides data about installed RPM Packages.\n====> * roce_scanner\n Detect active RoCE NICs on IBM Z machines.\n====> * read_openssh_config\n Collect information about the OpenSSH configuration.\n====> * persistentnetnames\n Get network interface information for physical ethernet interfaces of the original system.\n====> * root_scanner\n Scan the system root directory and produce a message containing\n====> * selinuxcontentscanner\n Scan the system for any SELinux customizations\n====> * scan_sap_hana\n Gathers information related to SAP HANA instances on the system.\n====> * check_custom_network_scripts\n Check the existence of custom network-scripts and warn user about possible\n====> * persistentnetnamesdisable\n Disable systemd-udevd persistent network naming on machine with single eth0 NIC\n====> * load_device_driver_deprecation_data\n Loads deprecation data for drivers and devices (PCI & CPU)\n====> * scan_custom_repofile\n Scan the custom /etc/leapp/files/leapp_upgrade_repositories.repo repo file.\n====> * trusted_gpg_keys_scanner\n Scan for trusted GPG keys.\n====> * distribution_signed_rpm_scanner\n Provide data about distribution signed & third-party RPM packages.\n====> * xfs_info_scanner\n This actor scans all mounted mountpoints for XFS information.\n====> * scan_default_initramfs\n Scan details of the default boot entry's initramfs image.\n====> * detect_kernel_drivers\n Matches all currently loaded kernel drivers against known deprecated and removed drivers.\n====> * used_repository_scanner\n Scan used enabled repositories\n====> * scan_dynamic_linker_configuration\n Scan the dynamic linker configuration and find modifications.\n====> * remove_obsolete_gpg_keys\n Remove obsoleted RPM GPG keys.\n====> * scan_source_kernel\n Scan the source system kernel.\n====> * repositories_blacklist\n Exclude target repositories provided by Red Hat without support.\n====> * get_installed_desktops\n Actor checks if kde or gnome desktop environments\n====> * scancpu\n Scan CPUs of the machine.\n====> * vdo_conversion_scanner\n Provides conversion info about VDO devices.\n====> * scan_hybrid_image_azure\n Check if the system is using Azure hybrid image.\n====> * satellite_upgrade_facts\n Report which Satellite packages require updates and how to handle PostgreSQL data\n====> * pci_devices_scanner\n Provides data about existing PCI Devices.\n====> * multipath_conf_read_8to9\n Read multipath configuration files and extract the necessary information\n====> * checkrhui\n Check if system is using RHUI infrastructure (on public cloud) and send messages to\n====> * biosdevname\n Enable biosdevname on the target RHEL system if all interfaces on the source RHEL\n====> * ipa_scanner\n Scan system for ipa-client and ipa-server status\n====> * rpm_transaction_config_tasks_collector\n Provides additional RPM transaction tasks from /etc/leapp/transaction.\n====> * pes_events_scanner\n Provides data about package events from Package Evolution Service.\n====> * satellite_upgrade_services\n Reconfigure Satellite services\n====> * setuptargetrepos\n Produces list of repositories that should be available to be used during IPU process.\n\n============================================================\n ERRORS \n============================================================\n\n2026-02-17 07:27:18.440450 [ERROR] Actor: scan_subscription_manager_info\nMessage: A subscription-manager command failed to execute\nSummary:\n Details: Command ['subscription-manager', 'release'] failed with exit code 1.\n Stderr: This system is not yet registered. Try 'subscription-manager register --help' for more information.\n Hint: Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\n Link: https://access.redhat.com/solutions/6138372\n\n============================================================\n END OF ERRORS \n============================================================\n\nDebug output written to /var/log/leapp/leapp-preupgrade.log\n\n============================================================\n REPORT OVERVIEW \n============================================================\n\nFollowing errors occurred and the upgrade cannot continue:\n 1. Actor: scan_subscription_manager_info\n Message: A subscription-manager command failed to execute\n\nHIGH and MEDIUM severity reports:\n 1. Packages available in excluded repositories will not be installed\n\nReports summary:\n Errors: 1\n Inhibitors: 0\n HIGH severity reports: 1\n MEDIUM severity reports: 0\n LOW severity reports: 0\n INFO severity reports: 1\n\nBefore continuing, review the full report below for details about discovered problems and possible remediation instructions:\n A report has been generated at /var/log/leapp/leapp-report.txt\n A report has been generated at /var/log/leapp/leapp-report.json\n\n============================================================\n END OF REPORT OVERVIEW \n============================================================\n\nAnswerfile has been generated at /var/log/leapp/answerfile", "stdout_lines": ["==> Processing phase `configuration_phase`", "====> * ipu_workflow_config", " IPU workflow config actor", "==> Processing phase `FactsCollection`", "====> * scan_target_os_image", " Scans the provided target OS ISO image to use as a content source for the IPU, if any.", "====> * scan_grub_config", " Scan grub configuration files for errors.", "====> * scancryptopolicies", " Scan information about system wide set crypto policies including:", "====> * udevadm_info", " Produces data exported by the \"udevadm info\" command.", "====> * transaction_workarounds", " Provides additional RPM transaction tasks based on bundled RPM packages.", "====> * repository_mapping", " Produces message containing repository mapping based on provided file.", "====> * scan_custom_modifications_actor", " Collects information about files in leapp directories that have been modified or newly added.", "====> * scandasd", " In case of s390x architecture, check whether DASD is used.", "====> * storage_scanner", " Provides data about storage settings.", "====> * network_manager_read_config", " Provides data about NetworkManager configuration.", "====> * ifcfg_scanner", " Scan ifcfg files with legacy network configuration", "====> * network_manager_connection_scanner", " Scan NetworkManager connection keyfiles", "====> * scan_files_for_target_userspace", " Scan the source system and identify files that will be copied into the target userspace when it is created.", "====> * scan_kernel_cmdline", " Scan the kernel command line of the booted system.", "====> * copy_dnf_conf_into_target_userspace", " Copy dnf.conf into target userspace", "====> * scanclienablerepo", " Produce CustomTargetRepository based on the LEAPP_ENABLE_REPOS in config.", "====> * firewalld_collect_used_object_names", " This actor reads firewalld's configuration and produces Model", "====> * system_facts", " Provides data about many facts from system.", "====> * scanblacklistca", " Scan the file system for distrusted CA's in the blacklist directory.", "====> * scan_pkg_manager", " Provides data about package manager (yum/dnf)", "====> * scanzfcp", " In case of s390x architecture, check whether ZFCP is used.", "====> * scan_source_files", " Scan files (explicitly specified) of the source system.", "====> * luks_scanner", " Provides data about active LUKS devices.", "====> * get_enabled_modules", " Provides data about which module streams are enabled on the source system.", "====> * scan_subscription_manager_info", " Scans the current system for subscription manager information", "====> * scan_source_boot_entry", " Scan the default boot entry of the source system.", "====> * firewalld_collect_global_config", " This actor reads firewalld's configuration and produces Model", "====> * scan_fips", " Determine whether the source system has FIPS enabled.", "====> * scan_grub_device_name", " Find the name of the block devices where GRUB is located", "====> * nis_scanner", " Collect information about the NIS packages configuration.", "====> * xorgdrvfacts8to9", " Check the journal logs for deprecated Xorg drivers.", "====> * scan_systemd_source", " Provides info about systemd on the source system", "====> * register_ruby_irb_adjustment", " Register a workaround to allow rubygem-irb's directory -> symlink conversion.", "====> * sssd_facts_8to9", " Check SSSD configuration for changes in RHEL9 and report them in model.", "====> * open_ssl_config_scanner", " Read an OpenSSL configuration file for further analysis.", "====> * scanmemory", " Scan Memory of the machine.", "====> * rpm_scanner", " Provides data about installed RPM Packages.", "====> * roce_scanner", " Detect active RoCE NICs on IBM Z machines.", "====> * read_openssh_config", " Collect information about the OpenSSH configuration.", "====> * persistentnetnames", " Get network interface information for physical ethernet interfaces of the original system.", "====> * root_scanner", " Scan the system root directory and produce a message containing", "====> * selinuxcontentscanner", " Scan the system for any SELinux customizations", "====> * scan_sap_hana", " Gathers information related to SAP HANA instances on the system.", "====> * check_custom_network_scripts", " Check the existence of custom network-scripts and warn user about possible", "====> * persistentnetnamesdisable", " Disable systemd-udevd persistent network naming on machine with single eth0 NIC", "====> * load_device_driver_deprecation_data", " Loads deprecation data for drivers and devices (PCI & CPU)", "====> * scan_custom_repofile", " Scan the custom /etc/leapp/files/leapp_upgrade_repositories.repo repo file.", "====> * trusted_gpg_keys_scanner", " Scan for trusted GPG keys.", "====> * distribution_signed_rpm_scanner", " Provide data about distribution signed & third-party RPM packages.", "====> * xfs_info_scanner", " This actor scans all mounted mountpoints for XFS information.", "====> * scan_default_initramfs", " Scan details of the default boot entry's initramfs image.", "====> * detect_kernel_drivers", " Matches all currently loaded kernel drivers against known deprecated and removed drivers.", "====> * used_repository_scanner", " Scan used enabled repositories", "====> * scan_dynamic_linker_configuration", " Scan the dynamic linker configuration and find modifications.", "====> * remove_obsolete_gpg_keys", " Remove obsoleted RPM GPG keys.", "====> * scan_source_kernel", " Scan the source system kernel.", "====> * repositories_blacklist", " Exclude target repositories provided by Red Hat without support.", "====> * get_installed_desktops", " Actor checks if kde or gnome desktop environments", "====> * scancpu", " Scan CPUs of the machine.", "====> * vdo_conversion_scanner", " Provides conversion info about VDO devices.", "====> * scan_hybrid_image_azure", " Check if the system is using Azure hybrid image.", "====> * satellite_upgrade_facts", " Report which Satellite packages require updates and how to handle PostgreSQL data", "====> * pci_devices_scanner", " Provides data about existing PCI Devices.", "====> * multipath_conf_read_8to9", " Read multipath configuration files and extract the necessary information", "====> * checkrhui", " Check if system is using RHUI infrastructure (on public cloud) and send messages to", "====> * biosdevname", " Enable biosdevname on the target RHEL system if all interfaces on the source RHEL", "====> * ipa_scanner", " Scan system for ipa-client and ipa-server status", "====> * rpm_transaction_config_tasks_collector", " Provides additional RPM transaction tasks from /etc/leapp/transaction.", "====> * pes_events_scanner", " Provides data about package events from Package Evolution Service.", "====> * satellite_upgrade_services", " Reconfigure Satellite services", "====> * setuptargetrepos", " Produces list of repositories that should be available to be used during IPU process.", "", "============================================================", " ERRORS ", "============================================================", "", "2026-02-17 07:27:18.440450 [ERROR] Actor: scan_subscription_manager_info", "Message: A subscription-manager command failed to execute", "Summary:", " Details: Command ['subscription-manager', 'release'] failed with exit code 1.", " Stderr: This system is not yet registered. Try 'subscription-manager register --help' for more information.", " Hint: Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.", " Link: https://access.redhat.com/solutions/6138372", "", "============================================================", " END OF ERRORS ", "============================================================", "", "Debug output written to /var/log/leapp/leapp-preupgrade.log", "", "============================================================", " REPORT OVERVIEW ", "============================================================", "", "Following errors occurred and the upgrade cannot continue:", " 1. Actor: scan_subscription_manager_info", " Message: A subscription-manager command failed to execute", "", "HIGH and MEDIUM severity reports:", " 1. Packages available in excluded repositories will not be installed", "", "Reports summary:", " Errors: 1", " Inhibitors: 0", " HIGH severity reports: 1", " MEDIUM severity reports: 0", " LOW severity reports: 0", " INFO severity reports: 1", "", "Before continuing, review the full report below for details about discovered problems and possible remediation instructions:", " A report has been generated at /var/log/leapp/leapp-report.txt", " A report has been generated at /var/log/leapp/leapp-report.json", "", "============================================================", " END OF REPORT OVERVIEW ", "============================================================", "", "Answerfile has been generated at /var/log/leapp/answerfile"]} TASK [analysis-leapp | Include custom_local_repos for local_repos_post_analysis] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:86 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [analysis-leapp | Restore original Satellite activation key] ************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:96 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : analysis-leapp | Include check-results-file.yml] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:107 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/check-results-file.yml for managed-node01 TASK [infra.leapp.analysis : check-results-file | Result file status] ********** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/check-results-file.yml:2 ok: [managed-node01] => {"changed": false, "stat": {"atime": 1771331259.0715117, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "f837348092b18f2a27bcc25c064bd72e50017924", "ctime": 1771331259.0715117, "dev": 51715, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 411041943, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1771331259.0715117, "nlink": 1, "path": "/var/log/leapp/leapp-report.txt", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 2702, "uid": 0, "version": "2078031566", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false}} TASK [infra.leapp.analysis : check-results-file | Check that result file exists] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/check-results-file.yml:7 ok: [managed-node01] => { "changed": false, "msg": "All assertions passed" } TASK [analysis-leapp | Run parse_leapp_report to check for inhibitors] ********* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:110 TASK [infra.leapp.common : parse_leapp_report | Default upgrade_inhibited to false] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:12 ok: [managed-node01] => {"ansible_facts": {"upgrade_inhibited": false}, "changed": false} TASK [infra.leapp.common : parse_leapp_report | Collect human readable report results] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:16 ok: [managed-node01] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false} TASK [infra.leapp.common : parse_leapp_report | Collect JSON report results] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:22 ok: [managed-node01] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false} TASK [infra.leapp.common : parse_leapp_report | Parse report results] ********** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:28 ok: [managed-node01] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false} TASK [infra.leapp.common : parse_leapp_report | Clear leapp_inhibitors] ******** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:35 ok: [managed-node01] => {"ansible_facts": {"leapp_inhibitors": []}, "changed": false} TASK [infra.leapp.common : parse_leapp_report | Check for inhibitors] ********** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:39 ok: [managed-node01] => (item={'audience': 'sysadmin', 'groups': ['error'], 'key': '7ec8269784db1bba2ac54ae438689ef397e16833', 'severity': 'high', 'summary': '{"details": "Command [\'subscription-manager\', \'release\'] failed with exit code 1.", "stderr": "This system is not yet registered. Try \'subscription-manager register --help\' for more information.\\n", "hint": "Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.", "link": "https://access.redhat.com/solutions/6138372"}', 'title': 'A subscription-manager command failed to execute', 'timeStamp': '2026-02-17T12:27:18.440660Z', 'hostname': 'managed-node01', 'actor': 'scan_subscription_manager_info', 'id': '46e191b52d96e170db8a1af978812c3b32a3b93ac5eccb3243b93323b4f1fd20'}) => {"ansible_facts": {"leapp_inhibitors": [{"actor": "scan_subscription_manager_info", "audience": "sysadmin", "groups": ["error"], "hostname": "managed-node01", "id": "46e191b52d96e170db8a1af978812c3b32a3b93ac5eccb3243b93323b4f1fd20", "key": "7ec8269784db1bba2ac54ae438689ef397e16833", "severity": "high", "summary": "{\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "timeStamp": "2026-02-17T12:27:18.440660Z", "title": "A subscription-manager command failed to execute"}], "upgrade_inhibited": true}, "ansible_loop_var": "item", "changed": false, "item": {"actor": "scan_subscription_manager_info", "audience": "sysadmin", "groups": ["error"], "hostname": "managed-node01", "id": "46e191b52d96e170db8a1af978812c3b32a3b93ac5eccb3243b93323b4f1fd20", "key": "7ec8269784db1bba2ac54ae438689ef397e16833", "severity": "high", "summary": "{\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "timeStamp": "2026-02-17T12:27:18.440660Z", "title": "A subscription-manager command failed to execute"}} skipping: [managed-node01] => (item={'audience': 'sysadmin', 'detail': {'remediations': [{'context': 'If some of excluded repositories are still required to be used during the upgrade, execute leapp with the --enablerepo option with the repoid of the repository required to be enabled as an argument (the option can be used multiple times).', 'type': 'hint'}]}, 'groups': ['repository', 'failure'], 'key': '1b9132cb2362ae7830e48eee7811be9527747de8', 'severity': 'info', 'summary': 'The following repositories are not supported by Red Hat and are excluded from the list of repositories used during the upgrade.\n- codeready-builder-for-rhel-9-ppc64le-eus-rpms\n- codeready-builder-beta-for-rhel-9-s390x-rpms\n- codeready-builder-for-rhel-9-aarch64-rpms\n- codeready-builder-for-rhel-9-x86_64-rpms\n- codeready-builder-for-rhel-9-rhui-rpms\n- codeready-builder-for-rhel-9-aarch64-eus-rpms\n- codeready-builder-for-rhel-9-x86_64-rhui-rpms\n- codeready-builder-for-rhel-9-ppc64le-rpms\n- codeready-builder-for-rhel-9-s390x-eus-rpms\n- codeready-builder-beta-for-rhel-9-aarch64-rpms\n- rhui-codeready-builder-for-rhel-9-aarch64-rhui-rpms\n- codeready-builder-for-rhel-9-x86_64-eus-rpms\n- crb\n- codeready-builder-for-rhel-9-s390x-rpms\n- codeready-builder-beta-for-rhel-9-ppc64le-rpms\n- codeready-builder-beta-for-rhel-9-x86_64-rpms\n- rhui-codeready-builder-for-rhel-9-x86_64-rhui-rpms', 'title': 'Excluded target system repositories', 'timeStamp': '2026-02-17T12:27:36.527513Z', 'hostname': 'managed-node01', 'actor': 'repositories_blacklist', 'id': '6273de9aa85c3dd95a92850660a4eb22fedb5eb64cc7a849f5226a729b90e64f'}) => {"ansible_loop_var": "item", "changed": false, "item": {"actor": "repositories_blacklist", "audience": "sysadmin", "detail": {"remediations": [{"context": "If some of excluded repositories are still required to be used during the upgrade, execute leapp with the --enablerepo option with the repoid of the repository required to be enabled as an argument (the option can be used multiple times).", "type": "hint"}]}, "groups": ["repository", "failure"], "hostname": "managed-node01", "id": "6273de9aa85c3dd95a92850660a4eb22fedb5eb64cc7a849f5226a729b90e64f", "key": "1b9132cb2362ae7830e48eee7811be9527747de8", "severity": "info", "summary": "The following repositories are not supported by Red Hat and are excluded from the list of repositories used during the upgrade.\n- codeready-builder-for-rhel-9-ppc64le-eus-rpms\n- codeready-builder-beta-for-rhel-9-s390x-rpms\n- codeready-builder-for-rhel-9-aarch64-rpms\n- codeready-builder-for-rhel-9-x86_64-rpms\n- codeready-builder-for-rhel-9-rhui-rpms\n- codeready-builder-for-rhel-9-aarch64-eus-rpms\n- codeready-builder-for-rhel-9-x86_64-rhui-rpms\n- codeready-builder-for-rhel-9-ppc64le-rpms\n- codeready-builder-for-rhel-9-s390x-eus-rpms\n- codeready-builder-beta-for-rhel-9-aarch64-rpms\n- rhui-codeready-builder-for-rhel-9-aarch64-rhui-rpms\n- codeready-builder-for-rhel-9-x86_64-eus-rpms\n- crb\n- codeready-builder-for-rhel-9-s390x-rpms\n- codeready-builder-beta-for-rhel-9-ppc64le-rpms\n- codeready-builder-beta-for-rhel-9-x86_64-rpms\n- rhui-codeready-builder-for-rhel-9-x86_64-rhui-rpms", "timeStamp": "2026-02-17T12:27:36.527513Z", "title": "Excluded target system repositories"}, "skip_reason": "Conditional result was False"} skipping: [managed-node01] => (item={'audience': 'sysadmin', 'detail': {'related_resources': [{'scheme': 'package', 'title': 'jitterentropy-devel'}]}, 'groups': ['repository'], 'key': '2437e204808f987477c0e9be8e4c95b3a87a9f3e', 'severity': 'high', 'summary': '1 packages will be skipped because they are available only in target system repositories that are intentionally excluded from the list of repositories used during the upgrade. See the report message titled "Excluded target system repositories" for details.\nThe list of these packages:\n- jitterentropy-devel (repoid: codeready-builder-for-rhel-9-x86_64-rpms)', 'title': 'Packages available in excluded repositories will not be installed', 'timeStamp': '2026-02-17T12:27:38.799817Z', 'hostname': 'managed-node01', 'actor': 'pes_events_scanner', 'id': '2b3a8b600eb5fe95b6573f3d788aa501fb5887d4ccff260176c910f1a9ca5579'}) => {"ansible_loop_var": "item", "changed": false, "item": {"actor": "pes_events_scanner", "audience": "sysadmin", "detail": {"related_resources": [{"scheme": "package", "title": "jitterentropy-devel"}]}, "groups": ["repository"], "hostname": "managed-node01", "id": "2b3a8b600eb5fe95b6573f3d788aa501fb5887d4ccff260176c910f1a9ca5579", "key": "2437e204808f987477c0e9be8e4c95b3a87a9f3e", "severity": "high", "summary": "1 packages will be skipped because they are available only in target system repositories that are intentionally excluded from the list of repositories used during the upgrade. See the report message titled \"Excluded target system repositories\" for details.\nThe list of these packages:\n- jitterentropy-devel (repoid: codeready-builder-for-rhel-9-x86_64-rpms)", "timeStamp": "2026-02-17T12:27:38.799817Z", "title": "Packages available in excluded repositories will not be installed"}, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : parse_leapp_report | Collect inhibitors] ************ task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:51 ok: [managed-node01] => {"changed": false, "cmd": ["awk", "/\\(inhibitor\\)/,/^-------/", "/var/log/leapp/leapp-report.txt"], "delta": "0:00:00.003848", "end": "2026-02-17 07:28:03.027323", "failed_when_result": false, "msg": "", "rc": 0, "start": "2026-02-17 07:28:03.023475", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [infra.leapp.common : parse_leapp_report | Collect high errors] *********** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:60 ok: [managed-node01] => {"changed": false, "cmd": ["awk", "/high \\(error\\)/,/^-------/", "/var/log/leapp/leapp-report.txt"], "delta": "0:00:00.004362", "end": "2026-02-17 07:28:03.365131", "failed_when_result": false, "msg": "", "rc": 0, "start": "2026-02-17 07:28:03.360769", "stderr": "", "stderr_lines": [], "stdout": "Risk Factor: high (error)\nTitle: A subscription-manager command failed to execute\nSummary: {\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}\nKey: 7ec8269784db1bba2ac54ae438689ef397e16833\n----------------------------------------", "stdout_lines": ["Risk Factor: high (error)", "Title: A subscription-manager command failed to execute", "Summary: {\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "Key: 7ec8269784db1bba2ac54ae438689ef397e16833", "----------------------------------------"]} TASK [infra.leapp.analysis : analysis-leapp | Copy reports to the controller] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:115 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/copy_reports_to_controller.yml for managed-node01 TASK [infra.leapp.analysis : copy_reports_to_controller | Ensure reports directory on controller] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/copy_reports_to_controller.yml:10 changed: [managed-node01 -> localhost] => {"changed": true, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tests/leapp_reports_2026-02-17_12-28-03", "secontext": "unconfined_u:object_r:admin_home_t:s0", "size": 6, "state": "directory", "uid": 0} TASK [infra.leapp.analysis : copy_reports_to_controller | Copy report files to the controller] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/copy_reports_to_controller.yml:18 changed: [managed-node01] => (item={'src': '/var/log/leapp/leapp-report.json', 'ext': 'json'}) => {"ansible_loop_var": "item", "changed": true, "checksum": "4b0af17a4e310bbe78ba1d195b62fd9dbf9053c2", "dest": "/root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tests/leapp_reports_2026-02-17_12-28-03/leapp-report.managed-node01.json/managed-node01/var/log/leapp/leapp-report.json", "item": {"ext": "json", "src": "/var/log/leapp/leapp-report.json"}, "md5sum": "f901c1d72363bc556ddef57e3a6742b3", "remote_checksum": "4b0af17a4e310bbe78ba1d195b62fd9dbf9053c2", "remote_md5sum": null} changed: [managed-node01] => (item={'src': '/var/log/leapp/leapp-report.txt', 'ext': 'txt'}) => {"ansible_loop_var": "item", "changed": true, "checksum": "f837348092b18f2a27bcc25c064bd72e50017924", "dest": "/root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tests/leapp_reports_2026-02-17_12-28-04/leapp-report.managed-node01.txt/managed-node01/var/log/leapp/leapp-report.txt", "item": {"ext": "txt", "src": "/var/log/leapp/leapp-report.txt"}, "md5sum": "00551d2ea0c55cd7fceafa883d813825", "remote_checksum": "f837348092b18f2a27bcc25c064bd72e50017924", "remote_md5sum": null} TASK [infra.leapp.analysis : analysis-leapp | Create remediation hostvars] ***** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:119 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/create_remediation_hostvars.yml for managed-node01 TASK [infra.leapp.analysis : create_remediation_hostvars | Clear remediation_todo] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/create_remediation_hostvars.yml:2 ok: [managed-node01] => {"ansible_facts": {"leapp_remediation_todo": []}, "changed": false} TASK [infra.leapp.analysis : create_remediation_hostvars | Map inhibitors to remediation_todo] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/create_remediation_hostvars.yml:6 failed: [managed-node01] (item=None) => {"ansible_loop_var": "inhibitor", "changed": false, "inhibitor": {"actor": "scan_subscription_manager_info", "audience": "sysadmin", "groups": ["error"], "hostname": "managed-node01", "id": "46e191b52d96e170db8a1af978812c3b32a3b93ac5eccb3243b93323b4f1fd20", "key": "7ec8269784db1bba2ac54ae438689ef397e16833", "severity": "high", "summary": "{\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "timeStamp": "2026-02-17T12:27:18.440660Z", "title": "A subscription-manager command failed to execute"}, "msg": "Failed to template loop_control.label: {{ __leapp_inhibitors_key_map[inhibitor_key]\n if inhibitor_key in __leapp_inhibitors_key_map\n else __leapp_inhibitors_title_map[inhibitor_title] }}: 'dict object' has no attribute 'A subscription-manager command failed to execute'. 'dict object' has no attribute 'A subscription-manager command failed to execute'. {{ __leapp_inhibitors_key_map[inhibitor_key]\n if inhibitor_key in __leapp_inhibitors_key_map\n else __leapp_inhibitors_title_map[inhibitor_title] }}: 'dict object' has no attribute 'A subscription-manager command failed to execute'. 'dict object' has no attribute 'A subscription-manager command failed to execute'", "skip_reason": "Conditional result was False"} TASK [Cleanup | Remove log files] ********************************************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tests/tests_default.yml:13 changed: [managed-node01] => {"changed": true, "cmd": "set -euxo pipefail\nrm -f /var/log/leapp/leapp-upgrade.log\nrm -f /var/log/ripu/ripu.log*\n", "delta": "0:00:00.004894", "end": "2026-02-17 07:28:04.778628", "msg": "", "rc": 0, "start": "2026-02-17 07:28:04.773734", "stderr": "+ rm -f /var/log/leapp/leapp-upgrade.log\n+ rm -f /var/log/ripu/ripu.log", "stderr_lines": ["+ rm -f /var/log/leapp/leapp-upgrade.log", "+ rm -f /var/log/ripu/ripu.log"], "stdout": "", "stdout_lines": []} PLAY RECAP ********************************************************************* managed-node01 : ok=30 changed=10 unreachable=0 failed=1 skipped=10 rescued=0 ignored=0 -- Logs begin at Tue 2026-02-17 07:23:35 EST, end at Tue 2026-02-17 07:28:05 EST. -- Feb 17 07:26:45 managed-node01 sshd[4578]: Accepted publickey for root from 10.31.9.232 port 38830 ssh2: ECDSA SHA256:dBMqjPTDcmFGQNgzSKN+9GYuDRCEPLXe5p12KGtoFN0 Feb 17 07:26:45 managed-node01 systemd[1]: Started Session 8 of user root. -- Subject: Unit session-8.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-8.scope has finished starting up. -- -- The start-up result is done. Feb 17 07:26:45 managed-node01 systemd-logind[602]: New session 8 of user root. -- Subject: A new session 8 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 8 has been created for the user root. -- -- The leading process of the session is 4578. Feb 17 07:26:45 managed-node01 sshd[4578]: pam_unix(sshd:session): session opened for user root by (uid=0) Feb 17 07:26:45 managed-node01 platform-python[4702]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Feb 17 07:26:46 managed-node01 platform-python[4836]: ansible-ansible.builtin.file Invoked with path=/var/log/ripu state=directory owner=root group=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 17 07:26:47 managed-node01 platform-python[4941]: ansible-ansible.builtin.stat Invoked with path=/var/log/ripu/ripu.log follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 17 07:26:47 managed-node01 platform-python[5046]: ansible-ansible.legacy.stat Invoked with path=/var/log/ripu/ripu.log follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 17 07:26:48 managed-node01 platform-python[5130]: ansible-ansible.legacy.copy Invoked with dest=/var/log/ripu/ripu.log owner=root group=root mode=0644 src=/root/.ansible/tmp/ansible-tmp-1771331207.3767955-5693-71201330276967/source _original_basename=tmp75iq283o follow=False checksum=59136dfcef86eeb56e4cbe8c783de491c0e0f501 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 17 07:26:48 managed-node01 platform-python[5237]: ansible-ansible.builtin.file Invoked with path=/etc/ansible/facts.d state=directory mode=0755 owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 17 07:26:48 managed-node01 platform-python[5342]: ansible-ansible.legacy.stat Invoked with path=/etc/ansible/facts.d/pre_ripu.fact follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 17 07:26:48 managed-node01 platform-python[5426]: ansible-ansible.legacy.copy Invoked with dest=/etc/ansible/facts.d/pre_ripu.fact mode=0644 owner=root group=root src=/root/.ansible/tmp/ansible-tmp-1771331208.4693317-5721-17461701937088/source _original_basename=tmpqz7foxp7 follow=False checksum=63939e8b278ceb3bc0722a169cbdd27ad14d45cc backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 17 07:26:49 managed-node01 platform-python[5533]: ansible-ansible.legacy.command Invoked with _raw_params=set -o pipefail; export PATH=$PATH; rpm -qa | grep -ve '[\.|+]el8' | grep -vE '^(gpg-pubkey|libmodulemd|katello-ca-consumer)' | sort _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 17 07:26:50 managed-node01 platform-python[5643]: ansible-ansible.legacy.stat Invoked with path=/etc/ansible/facts.d/non_rhel_packages.fact follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 17 07:26:50 managed-node01 platform-python[5727]: ansible-ansible.legacy.copy Invoked with dest=/etc/ansible/facts.d/non_rhel_packages.fact mode=0644 owner=root group=root src=/root/.ansible/tmp/ansible-tmp-1771331210.3934922-5754-242075813388270/source _original_basename=tmpwqk6z_9e follow=False checksum=97d170e1550eee4afc0af065b78cda302a97674c backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 17 07:26:51 managed-node01 platform-python[5834]: ansible-ansible.legacy.dnf Invoked with name=['leapp-upgrade'] enablerepo=[] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 17 07:26:57 managed-node01 dbus-daemon[606]: [system] Reloaded configuration Feb 17 07:26:57 managed-node01 dbus-daemon[606]: [system] Reloaded configuration Feb 17 07:26:57 managed-node01 dbus-daemon[606]: [system] Reloaded configuration Feb 17 07:26:57 managed-node01 dbus-daemon[606]: [system] Reloaded configuration Feb 17 07:26:57 managed-node01 dbus-daemon[606]: [system] Reloaded configuration Feb 17 07:26:57 managed-node01 dbus-daemon[606]: [system] Reloaded configuration Feb 17 07:26:58 managed-node01 systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. -- Subject: Unit run-r8ab01f8c21f34ceba3c6f2197ef69c73.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit run-r8ab01f8c21f34ceba3c6f2197ef69c73.service has finished starting up. -- -- The start-up result is done. Feb 17 07:26:58 managed-node01 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Feb 17 07:26:58 managed-node01 systemd[1]: Starting man-db-cache-update.service... -- Subject: Unit man-db-cache-update.service has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit man-db-cache-update.service has begun starting up. Feb 17 07:26:58 managed-node01 systemd[1]: Reloading. Feb 17 07:26:59 managed-node01 platform-python[6409]: ansible-ansible.builtin.file Invoked with path=/var/log/leapp state=directory owner=root group=root mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 17 07:27:00 managed-node01 systemd[1]: man-db-cache-update.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit man-db-cache-update.service has successfully entered the 'dead' state. Feb 17 07:27:00 managed-node01 systemd[1]: Started man-db-cache-update.service. -- Subject: Unit man-db-cache-update.service has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit man-db-cache-update.service has finished starting up. -- -- The start-up result is done. Feb 17 07:27:00 managed-node01 systemd[1]: run-r8ab01f8c21f34ceba3c6f2197ef69c73.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit run-r8ab01f8c21f34ceba3c6f2197ef69c73.service has successfully entered the 'dead' state. Feb 17 07:27:00 managed-node01 ansible-async_wrapper.py[6621]: Invoked with j3410284144 7200 /root/.ansible/tmp/ansible-tmp-1771331220.0591471-5834-207301654554008/AnsiballZ_command.py _ Feb 17 07:27:00 managed-node01 ansible-async_wrapper.py[6625]: Starting module and watcher Feb 17 07:27:00 managed-node01 ansible-async_wrapper.py[6625]: Start watching 6626 (7200) Feb 17 07:27:00 managed-node01 ansible-async_wrapper.py[6626]: Start module (6626) Feb 17 07:27:00 managed-node01 ansible-async_wrapper.py[6621]: Return async_wrapper task started. Feb 17 07:27:00 managed-node01 platform-python[6627]: ansible-ansible.legacy.command Invoked with executable=/bin/bash _raw_params=set -o pipefail; export PATH=$PATH; ulimit -n 16384; leapp preupgrade --report-schema=1.2.0 2>&1 | tee -a /var/log/ripu/ripu.log _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None creates=None removes=None stdin=None Feb 17 07:27:05 managed-node01 ansible-async_wrapper.py[6625]: 6626 still running (7200) Feb 17 07:27:10 managed-node01 ansible-async_wrapper.py[6625]: 6626 still running (7195) Feb 17 07:27:13 managed-node01 systemd[1]: proc-sys-fs-binfmt_misc.automount: Got automount request for /proc/sys/fs/binfmt_misc, triggered by 8004 (sysctl) Feb 17 07:27:13 managed-node01 systemd[1]: Mounting Arbitrary Executable File Formats File System... -- Subject: Unit proc-sys-fs-binfmt_misc.mount has begun start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit proc-sys-fs-binfmt_misc.mount has begun starting up. Feb 17 07:27:13 managed-node01 systemd[1]: Mounted Arbitrary Executable File Formats File System. -- Subject: Unit proc-sys-fs-binfmt_misc.mount has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit proc-sys-fs-binfmt_misc.mount has finished starting up. -- -- The start-up result is done. Feb 17 07:27:15 managed-node01 ansible-async_wrapper.py[6625]: 6626 still running (7190) Feb 17 07:27:20 managed-node01 ansible-async_wrapper.py[6625]: 6626 still running (7185) Feb 17 07:27:25 managed-node01 ansible-async_wrapper.py[6625]: 6626 still running (7180) Feb 17 07:27:30 managed-node01 ansible-async_wrapper.py[6625]: 6626 still running (7175) Feb 17 07:27:35 managed-node01 ansible-async_wrapper.py[6625]: 6626 still running (7170) Feb 17 07:27:39 managed-node01 ansible-async_wrapper.py[6626]: Module complete (6626) Feb 17 07:27:40 managed-node01 ansible-async_wrapper.py[6625]: Done in kid B. Feb 17 07:27:59 managed-node01 sshd[4581]: Received disconnect from 10.31.9.232 port 38830:11: disconnected by user Feb 17 07:27:59 managed-node01 sshd[4581]: Disconnected from user root 10.31.9.232 port 38830 Feb 17 07:27:59 managed-node01 sshd[4578]: pam_unix(sshd:session): session closed for user root Feb 17 07:27:59 managed-node01 systemd[1]: session-8.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit session-8.scope has successfully entered the 'dead' state. Feb 17 07:27:59 managed-node01 systemd-logind[602]: Session 8 logged out. Waiting for processes to exit. Feb 17 07:27:59 managed-node01 systemd-logind[602]: Removed session 8. -- Subject: Session 8 has been terminated -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A session with the ID 8 has been terminated. Feb 17 07:28:00 managed-node01 sshd[10823]: Accepted publickey for root from 10.31.9.232 port 57910 ssh2: ECDSA SHA256:dBMqjPTDcmFGQNgzSKN+9GYuDRCEPLXe5p12KGtoFN0 Feb 17 07:28:00 managed-node01 systemd-logind[602]: New session 9 of user root. -- Subject: A new session 9 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 9 has been created for the user root. -- -- The leading process of the session is 10823. Feb 17 07:28:00 managed-node01 systemd[1]: Started Session 9 of user root. -- Subject: Unit session-9.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-9.scope has finished starting up. -- -- The start-up result is done. Feb 17 07:28:00 managed-node01 sshd[10823]: pam_unix(sshd:session): session opened for user root by (uid=0) Feb 17 07:28:01 managed-node01 platform-python[10929]: ansible-ansible.legacy.async_status Invoked with jid=j3410284144.6621 mode=status _async_dir=/root/.ansible_async Feb 17 07:28:01 managed-node01 platform-python[10997]: ansible-ansible.legacy.async_status Invoked with jid=j3410284144.6621 mode=cleanup _async_dir=/root/.ansible_async Feb 17 07:28:01 managed-node01 platform-python[11102]: ansible-ansible.builtin.stat Invoked with path=/var/log/leapp/leapp-report.txt follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 17 07:28:03 managed-node01 platform-python[11419]: ansible-ansible.legacy.command Invoked with _raw_params=awk '/\(inhibitor\)/,/^-------/' /var/log/leapp/leapp-report.txt _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 17 07:28:03 managed-node01 platform-python[11525]: ansible-ansible.legacy.command Invoked with _raw_params=awk '/high \(error\)/,/^-------/' /var/log/leapp/leapp-report.txt _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 17 07:28:03 managed-node01 platform-python[11631]: ansible-ansible.legacy.stat Invoked with path=/var/log/leapp/leapp-report.json follow=True get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 17 07:28:04 managed-node01 platform-python[11753]: ansible-ansible.legacy.stat Invoked with path=/var/log/leapp/leapp-report.txt follow=True get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 17 07:28:04 managed-node01 platform-python[11875]: ansible-ansible.legacy.command Invoked with executable=/bin/bash _raw_params=set -euxo pipefail rm -f /var/log/leapp/leapp-upgrade.log rm -f /var/log/ripu/ripu.log* _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None creates=None removes=None stdin=None Feb 17 07:28:04 managed-node01 sshd[11895]: Accepted publickey for root from 10.31.9.232 port 57914 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Feb 17 07:28:04 managed-node01 systemd-logind[602]: New session 10 of user root. -- Subject: A new session 10 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 10 has been created for the user root. -- -- The leading process of the session is 11895. Feb 17 07:28:04 managed-node01 systemd[1]: Started Session 10 of user root. -- Subject: Unit session-10.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-10.scope has finished starting up. -- -- The start-up result is done. Feb 17 07:28:04 managed-node01 sshd[11895]: pam_unix(sshd:session): session opened for user root by (uid=0) Feb 17 07:28:05 managed-node01 sshd[11898]: Received disconnect from 10.31.9.232 port 57914:11: disconnected by user Feb 17 07:28:05 managed-node01 sshd[11898]: Disconnected from user root 10.31.9.232 port 57914 Feb 17 07:28:05 managed-node01 sshd[11895]: pam_unix(sshd:session): session closed for user root Feb 17 07:28:05 managed-node01 systemd[1]: session-10.scope: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- The unit session-10.scope has successfully entered the 'dead' state. Feb 17 07:28:05 managed-node01 systemd-logind[602]: Session 10 logged out. Waiting for processes to exit. Feb 17 07:28:05 managed-node01 systemd-logind[602]: Removed session 10. -- Subject: Session 10 has been terminated -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A session with the ID 10 has been terminated. Feb 17 07:28:05 managed-node01 sshd[11916]: Accepted publickey for root from 10.31.9.232 port 57928 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Feb 17 07:28:05 managed-node01 systemd[1]: Started Session 11 of user root. -- Subject: Unit session-11.scope has finished start-up -- Defined-By: systemd -- Support: https://access.redhat.com/support -- -- Unit session-11.scope has finished starting up. -- -- The start-up result is done. Feb 17 07:28:05 managed-node01 systemd-logind[602]: New session 11 of user root. -- Subject: A new session 11 has been created for user root -- Defined-By: systemd -- Support: https://access.redhat.com/support -- Documentation: https://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 11 has been created for the user root. -- -- The leading process of the session is 11916. Feb 17 07:28:05 managed-node01 sshd[11916]: pam_unix(sshd:session): session opened for user root by (uid=0)