[WARNING]: Collection infra.leapp does not support Ansible version 2.14.18 [WARNING]: running playbook inside collection infra.leapp ansible-playbook [core 2.14.18] config file = /etc/ansible/ansible.cfg configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.9/site-packages/ansible ansible collection location = /root/.ansible/collections:/usr/share/ansible/collections executable location = /usr/bin/ansible-playbook python version = 3.9.25 (main, Jan 14 2026, 00:00:00) [GCC 11.5.0 20240719 (Red Hat 11.5.0-11)] (/usr/bin/python3) jinja version = 3.1.2 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_default.yml **************************************************** 1 plays in /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tests/tests_default.yml PLAY [Test] ******************************************************************** TASK [Gathering Facts] ********************************************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tests/tests_default.yml:2 ok: [managed-node01] TASK [Test | Run role analysis] ************************************************ task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tests/tests_default.yml:9 TASK [infra.leapp.common : Log directory exists] ******************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:3 changed: [managed-node01] => {"changed": true, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/var/log/ripu", "secontext": "unconfined_u:object_r:var_log_t:s0", "size": 6, "state": "directory", "uid": 0} TASK [infra.leapp.common : Check for existing log file] ************************ task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:11 ok: [managed-node01] => {"changed": false, "stat": {"exists": false}} TASK [infra.leapp.common : Fail if log file already exists] ******************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:16 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : Create new log file] ******************************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:23 NOTIFIED HANDLER infra.leapp.common : Check for log file for managed-node01 NOTIFIED HANDLER infra.leapp.common : Add end time to log file for managed-node01 NOTIFIED HANDLER infra.leapp.common : Slurp ripu.log file for managed-node01 NOTIFIED HANDLER infra.leapp.common : Decode ripu.log file for managed-node01 NOTIFIED HANDLER infra.leapp.common : Rename log file for managed-node01 changed: [managed-node01] => {"changed": true, "checksum": "d07f0c7b48599f0c4b54e09ada45d9885b771962", "dest": "/var/log/ripu/ripu.log", "gid": 0, "group": "root", "md5sum": "abc4b44f2ad1cd736fc7c9b4f0f319dd", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:var_log_t:s0", "size": 61, "src": "/root/.ansible/tmp/ansible-tmp-1771334580.6720881-5680-26692498008197/source", "state": "file", "uid": 0} TASK [infra.leapp.common : /etc/ansible/facts.d directory exists] ************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:35 changed: [managed-node01] => {"changed": true, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/ansible/facts.d", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0} TASK [infra.leapp.common : Capture current ansible_facts for validation after upgrade] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:43 changed: [managed-node01] => {"changed": true, "checksum": "1b348fd1227c8d47b88a0a2c38947fb1940bc624", "dest": "/etc/ansible/facts.d/pre_ripu.fact", "gid": 0, "group": "root", "md5sum": "e3d1092c26b01aa8f1bc48952512aa76", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 14666, "src": "/root/.ansible/tmp/ansible-tmp-1771334581.8351882-5710-132141815384946/source", "state": "file", "uid": 0} TASK [infra.leapp.common : Capture a list of non-rhel versioned packages] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:51 ok: [managed-node01] => {"changed": false, "cmd": "set -o pipefail; export PATH=$PATH; rpm -qa | grep -ve '[\\.|+]el9' | grep -vE '^(gpg-pubkey|libmodulemd|katello-ca-consumer)' | sort", "delta": "0:00:00.180914", "end": "2026-02-17 08:23:03.042344", "failed_when_result": false, "msg": "non-zero return code", "rc": 1, "start": "2026-02-17 08:23:02.861430", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [infra.leapp.common : Create fact with the non-rhel versioned packages list] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:65 ok: [managed-node01] => {"ansible_facts": {"non_rhel_packages": []}, "changed": false} TASK [infra.leapp.common : Capture the list of non-rhel versioned packages in a separate fact file] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/main.yml:69 changed: [managed-node01] => {"changed": true, "checksum": "97d170e1550eee4afc0af065b78cda302a97674c", "dest": "/etc/ansible/facts.d/non_rhel_packages.fact", "gid": 0, "group": "root", "md5sum": "d751713988987e9331980363e24189ce", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 2, "src": "/root/.ansible/tmp/ansible-tmp-1771334583.1583416-5741-258761498105593/source", "state": "file", "uid": 0} TASK [infra.leapp.analysis : Include tasks for preupg assistant analysis] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/main.yml:9 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : Include tasks for leapp preupgrade analysis] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/main.yml:13 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml for managed-node01 TASK [analysis-leapp | Register with Satellite activation key] ***************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:2 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [analysis-leapp | Include custom_local_repos for local_repos_pre_leapp] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:11 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : analysis-leapp | Install packages for preupgrade analysis on RHEL 7] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:19 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : analysis-leapp | Install packages for preupgrade analysis on RHEL 8] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:26 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : analysis-leapp | Install packages for preupgrade analysis on RHEL 9] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:33 changed: [managed-node01] => {"changed": true, "msg": "", "rc": 0, "results": ["Installed: python3-pyudev-0.22.0-6.el9.noarch", "Installed: systemd-container-252-55.el9_7.7.x86_64", "Installed: python3-leapp-0.20.0-1.el9.noarch", "Installed: policycoreutils-python-utils-3.6-3.el9.noarch", "Installed: pciutils-3.7.0-7.el9.x86_64", "Installed: libdb-utils-5.3.28-57.el9_6.x86_64", "Installed: leapp-0.20.0-1.el9.noarch", "Installed: leapp-deps-0.20.0-1.el9.noarch", "Installed: leapp-upgrade-el9toel10-0.23.0-1.el9.noarch", "Installed: leapp-upgrade-el9toel10-deps-0.23.0-1.el9.noarch"]} TASK [infra.leapp.analysis : analysis-leapp | Ensure leapp log directory exists] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:40 ok: [managed-node01] => {"changed": false, "gid": 0, "group": "root", "mode": "0700", "owner": "root", "path": "/var/log/leapp", "secontext": "system_u:object_r:var_log_t:s0", "size": 6, "state": "directory", "uid": 0} TASK [infra.leapp.analysis : analysis-leapp | Populate leapp_answers file] ***** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:48 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [analysis-leapp | Create /etc/leapp/files/leapp_upgrade_repositories.repo] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:57 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : analysis-leapp | Leapp preupgrade report] ********* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:68 ASYNC FAILED on managed-node01: jid=j108107829421.7347 changed: [managed-node01] => {"ansible_job_id": "j108107829421.7347", "changed": true, "cmd": "set -o pipefail; export PATH=$PATH; ulimit -n 16384; leapp preupgrade --report-schema=1.2.0 2>&1 | tee -a /var/log/ripu/ripu.log\n", "delta": "0:00:29.819686", "end": "2026-02-17 08:23:45.842745", "failed_when_result": false, "finished": 1, "msg": "non-zero return code", "rc": 1, "results_file": "/root/.ansible_async/j108107829421.7347", "start": "2026-02-17 08:23:16.023059", "started": 1, "stderr": "", "stderr_lines": [], "stdout": "==> Processing phase `configuration_phase`\n====> * ipu_workflow_config\n IPU workflow config actor\n==> Processing phase `FactsCollection`\n====> * root_scanner\n Scan the system root directory and produce a message containing\n====> * scan_custom_modifications_actor\n Collects information about files in leapp directories that have been modified or newly added.\n====> * scan_grub_device_name\n Find the name of the block devices where GRUB is located\n====> * scancryptopolicies\n Scan information about system wide set crypto policies including:\n====> * transaction_workarounds\n Provides additional RPM transaction tasks based on bundled RPM packages.\n====> * scanclienablerepo\n Produce CustomTargetRepository based on the LEAPP_ENABLE_REPOS in config.\n====> * scan_custom_repofile\n Scan the custom /etc/leapp/files/leapp_upgrade_repositories.repo repo file.\n====> * scan_systemd_source\n Provides info about systemd on the source system\n====> * satellite_upgrade_services\n Reconfigure Satellite services\n====> * scandasd\n In case of s390x architecture, check whether DASD is used.\n====> * scan_files_for_target_userspace\n Scan the source system and identify files that will be copied into the target userspace when it is created.\n====> * scanmemory\n Scan Memory of the machine.\n====> * ifcfg_scanner\n Scan ifcfg files with legacy network configuration\n====> * scan_grub_config\n Scan grub configuration files for errors.\n====> * scan_sap_hana\n Gathers information related to SAP HANA instances on the system.\n====> * get_enabled_modules\n Provides data about which module streams are enabled on the source system.\n====> * scan_kernel_cmdline\n Scan the kernel command line of the booted system.\n====> * migrate_rpm_db\n Register a workaround to migrate RPM DB during the upgrade.\n====> * scan_source_files\n Scan files (explicitly specified) of the source system.\n====> * scan_pkg_manager\n Provides data about package manager (yum/dnf)\n====> * scan_subscription_manager_info\n Scans the current system for subscription manager information\n====> * register_ruby_irb_adjustment\n Register a workaround to allow rubygem-irb's symlink -> directory conversion.\n====> * scan_target_os_image\n Scans the provided target OS ISO image to use as a content source for the IPU, if any.\n====> * repository_mapping\n Produces message containing repository mapping based on provided file.\n====> * scanzfcp\n In case of s390x architecture, check whether ZFCP is used.\n====> * network_manager_read_config\n Provides data about NetworkManager configuration.\n====> * scan_pam_user_db\n Scan the PAM service folder for the location of pam_userdb databases\n====> * udevadm_info\n Produces data exported by the \"udevadm info\" command.\n====> * load_device_driver_deprecation_data\n Loads deprecation data for drivers and devices (PCI & CPU)\n====> * sssd_facts\n Check SSSD configuration for changes in RHEL10 and report them in model.\n====> * copy_dnf_conf_into_target_userspace\n Copy dnf.conf into target userspace\n====> * storage_scanner\n Provides data about storage settings.\n====> * scan_source_boot_entry\n Scan the default boot entry of the source system.\n====> * read_openssh_config\n Collect information about the OpenSSH configuration.\n====> * open_ssl_config_scanner\n Read an OpenSSL configuration file for further analysis.\n====> * rpm_scanner\n Provides data about installed RPM Packages.\n====> * system_facts\n Provides data about many facts from system.\n====> * xfs_info_scanner\n This actor scans all mounted mountpoints for XFS information.\n====> * persistentnetnames\n Get network interface information for physical ethernet interfaces of the original system.\n====> * scan_hybrid_image_azure\n Check if the system is using Azure hybrid image.\n====> * scan_default_initramfs\n Scan details of the default boot entry's initramfs image.\n====> * detect_kernel_drivers\n Matches all currently loaded kernel drivers against known deprecated and removed drivers.\n====> * persistentnetnamesdisable\n Disable systemd-udevd persistent network naming on machine with single eth0 NIC\n====> * scancpu\n Scan CPUs of the machine.\n====> * scan_fips\n Determine whether the source system has FIPS enabled.\n====> * get_installed_desktops\n Actor checks if kde or gnome desktop environments\n====> * luks_scanner\n Provides data about active LUKS devices.\n====> * biosdevname\n Enable biosdevname on the target RHEL system if all interfaces on the source RHEL\n====> * pci_devices_scanner\n Provides data about existing PCI Devices.\n====> * distribution_signed_rpm_scanner\n Provide data about distribution signed & third-party RPM packages.\n====> * trusted_gpg_keys_scanner\n Scan for trusted GPG keys.\n====> * used_repository_scanner\n Scan used enabled repositories\n====> * selinuxcontentscanner\n Scan the system for any SELinux customizations\n====> * remove_obsolete_gpg_keys\n Remove obsoleted RPM GPG keys.\n====> * rpm_transaction_config_tasks_collector\n Provides additional RPM transaction tasks from /etc/leapp/transaction.\n====> * scan_mysql\n Actor checking for presence of MySQL installation.\n====> * repositories_blacklist\n Exclude target repositories provided by Red Hat without support.\n====> * checkrhui\n Check if system is using RHUI infrastructure (on public cloud) and send messages to\n====> * pes_events_scanner\n Provides data about package events from Package Evolution Service.\n====> * scan_dynamic_linker_configuration\n Scan the dynamic linker configuration and find modifications.\n====> * ipa_scanner\n Scan system for ipa-client and ipa-server status\n====> * scan_source_kernel\n Scan the source system kernel.\n====> * setuptargetrepos\n Produces list of repositories that should be available to be used during IPU process.\n\n============================================================\n ERRORS \n============================================================\n\n2026-02-17 08:23:27.285528 [ERROR] Actor: scan_subscription_manager_info\nMessage: A subscription-manager command failed to execute\nSummary:\n Details: Command ['subscription-manager', 'release'] failed with exit code 1.\n Stderr: This system is not yet registered. Try 'subscription-manager register --help' for more information.\n Hint: Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\n Link: https://access.redhat.com/solutions/6138372\n\n============================================================\n END OF ERRORS \n============================================================\n\nDebug output written to /var/log/leapp/leapp-preupgrade.log\n\n============================================================\n REPORT OVERVIEW \n============================================================\n\nFollowing errors occurred and the upgrade cannot continue:\n 1. Actor: scan_subscription_manager_info\n Message: A subscription-manager command failed to execute\n\nReports summary:\n Errors: 1\n Inhibitors: 0\n HIGH severity reports: 0\n MEDIUM severity reports: 0\n LOW severity reports: 0\n INFO severity reports: 1\n\nBefore continuing, review the full report below for details about discovered problems and possible remediation instructions:\n A report has been generated at /var/log/leapp/leapp-report.txt\n A report has been generated at /var/log/leapp/leapp-report.json\n\n============================================================\n END OF REPORT OVERVIEW \n============================================================\n\nAnswerfile has been generated at /var/log/leapp/answerfile", "stdout_lines": ["==> Processing phase `configuration_phase`", "====> * ipu_workflow_config", " IPU workflow config actor", "==> Processing phase `FactsCollection`", "====> * root_scanner", " Scan the system root directory and produce a message containing", "====> * scan_custom_modifications_actor", " Collects information about files in leapp directories that have been modified or newly added.", "====> * scan_grub_device_name", " Find the name of the block devices where GRUB is located", "====> * scancryptopolicies", " Scan information about system wide set crypto policies including:", "====> * transaction_workarounds", " Provides additional RPM transaction tasks based on bundled RPM packages.", "====> * scanclienablerepo", " Produce CustomTargetRepository based on the LEAPP_ENABLE_REPOS in config.", "====> * scan_custom_repofile", " Scan the custom /etc/leapp/files/leapp_upgrade_repositories.repo repo file.", "====> * scan_systemd_source", " Provides info about systemd on the source system", "====> * satellite_upgrade_services", " Reconfigure Satellite services", "====> * scandasd", " In case of s390x architecture, check whether DASD is used.", "====> * scan_files_for_target_userspace", " Scan the source system and identify files that will be copied into the target userspace when it is created.", "====> * scanmemory", " Scan Memory of the machine.", "====> * ifcfg_scanner", " Scan ifcfg files with legacy network configuration", "====> * scan_grub_config", " Scan grub configuration files for errors.", "====> * scan_sap_hana", " Gathers information related to SAP HANA instances on the system.", "====> * get_enabled_modules", " Provides data about which module streams are enabled on the source system.", "====> * scan_kernel_cmdline", " Scan the kernel command line of the booted system.", "====> * migrate_rpm_db", " Register a workaround to migrate RPM DB during the upgrade.", "====> * scan_source_files", " Scan files (explicitly specified) of the source system.", "====> * scan_pkg_manager", " Provides data about package manager (yum/dnf)", "====> * scan_subscription_manager_info", " Scans the current system for subscription manager information", "====> * register_ruby_irb_adjustment", " Register a workaround to allow rubygem-irb's symlink -> directory conversion.", "====> * scan_target_os_image", " Scans the provided target OS ISO image to use as a content source for the IPU, if any.", "====> * repository_mapping", " Produces message containing repository mapping based on provided file.", "====> * scanzfcp", " In case of s390x architecture, check whether ZFCP is used.", "====> * network_manager_read_config", " Provides data about NetworkManager configuration.", "====> * scan_pam_user_db", " Scan the PAM service folder for the location of pam_userdb databases", "====> * udevadm_info", " Produces data exported by the \"udevadm info\" command.", "====> * load_device_driver_deprecation_data", " Loads deprecation data for drivers and devices (PCI & CPU)", "====> * sssd_facts", " Check SSSD configuration for changes in RHEL10 and report them in model.", "====> * copy_dnf_conf_into_target_userspace", " Copy dnf.conf into target userspace", "====> * storage_scanner", " Provides data about storage settings.", "====> * scan_source_boot_entry", " Scan the default boot entry of the source system.", "====> * read_openssh_config", " Collect information about the OpenSSH configuration.", "====> * open_ssl_config_scanner", " Read an OpenSSL configuration file for further analysis.", "====> * rpm_scanner", " Provides data about installed RPM Packages.", "====> * system_facts", " Provides data about many facts from system.", "====> * xfs_info_scanner", " This actor scans all mounted mountpoints for XFS information.", "====> * persistentnetnames", " Get network interface information for physical ethernet interfaces of the original system.", "====> * scan_hybrid_image_azure", " Check if the system is using Azure hybrid image.", "====> * scan_default_initramfs", " Scan details of the default boot entry's initramfs image.", "====> * detect_kernel_drivers", " Matches all currently loaded kernel drivers against known deprecated and removed drivers.", "====> * persistentnetnamesdisable", " Disable systemd-udevd persistent network naming on machine with single eth0 NIC", "====> * scancpu", " Scan CPUs of the machine.", "====> * scan_fips", " Determine whether the source system has FIPS enabled.", "====> * get_installed_desktops", " Actor checks if kde or gnome desktop environments", "====> * luks_scanner", " Provides data about active LUKS devices.", "====> * biosdevname", " Enable biosdevname on the target RHEL system if all interfaces on the source RHEL", "====> * pci_devices_scanner", " Provides data about existing PCI Devices.", "====> * distribution_signed_rpm_scanner", " Provide data about distribution signed & third-party RPM packages.", "====> * trusted_gpg_keys_scanner", " Scan for trusted GPG keys.", "====> * used_repository_scanner", " Scan used enabled repositories", "====> * selinuxcontentscanner", " Scan the system for any SELinux customizations", "====> * remove_obsolete_gpg_keys", " Remove obsoleted RPM GPG keys.", "====> * rpm_transaction_config_tasks_collector", " Provides additional RPM transaction tasks from /etc/leapp/transaction.", "====> * scan_mysql", " Actor checking for presence of MySQL installation.", "====> * repositories_blacklist", " Exclude target repositories provided by Red Hat without support.", "====> * checkrhui", " Check if system is using RHUI infrastructure (on public cloud) and send messages to", "====> * pes_events_scanner", " Provides data about package events from Package Evolution Service.", "====> * scan_dynamic_linker_configuration", " Scan the dynamic linker configuration and find modifications.", "====> * ipa_scanner", " Scan system for ipa-client and ipa-server status", "====> * scan_source_kernel", " Scan the source system kernel.", "====> * setuptargetrepos", " Produces list of repositories that should be available to be used during IPU process.", "", "============================================================", " ERRORS ", "============================================================", "", "2026-02-17 08:23:27.285528 [ERROR] Actor: scan_subscription_manager_info", "Message: A subscription-manager command failed to execute", "Summary:", " Details: Command ['subscription-manager', 'release'] failed with exit code 1.", " Stderr: This system is not yet registered. Try 'subscription-manager register --help' for more information.", " Hint: Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.", " Link: https://access.redhat.com/solutions/6138372", "", "============================================================", " END OF ERRORS ", "============================================================", "", "Debug output written to /var/log/leapp/leapp-preupgrade.log", "", "============================================================", " REPORT OVERVIEW ", "============================================================", "", "Following errors occurred and the upgrade cannot continue:", " 1. Actor: scan_subscription_manager_info", " Message: A subscription-manager command failed to execute", "", "Reports summary:", " Errors: 1", " Inhibitors: 0", " HIGH severity reports: 0", " MEDIUM severity reports: 0", " LOW severity reports: 0", " INFO severity reports: 1", "", "Before continuing, review the full report below for details about discovered problems and possible remediation instructions:", " A report has been generated at /var/log/leapp/leapp-report.txt", " A report has been generated at /var/log/leapp/leapp-report.json", "", "============================================================", " END OF REPORT OVERVIEW ", "============================================================", "", "Answerfile has been generated at /var/log/leapp/answerfile"]} TASK [analysis-leapp | Include custom_local_repos for local_repos_post_analysis] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:86 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [analysis-leapp | Restore original Satellite activation key] ************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:96 skipping: [managed-node01] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : analysis-leapp | Include check-results-file.yml] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:107 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/check-results-file.yml for managed-node01 TASK [infra.leapp.analysis : check-results-file | Result file status] ********** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/check-results-file.yml:2 ok: [managed-node01] => {"changed": false, "stat": {"atime": 1771334625.3888032, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "5b8d6f66575fd6bcfdf679dd6d49c929c7412053", "ctime": 1771334625.3888032, "dev": 51716, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 310378654, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1771334625.3888032, "nlink": 1, "path": "/var/log/leapp/leapp-report.txt", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 2123, "uid": 0, "version": "1709911325", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false}} TASK [infra.leapp.analysis : check-results-file | Check that result file exists] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/check-results-file.yml:7 ok: [managed-node01] => { "changed": false, "msg": "All assertions passed" } TASK [analysis-leapp | Run parse_leapp_report to check for inhibitors] ********* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:110 TASK [infra.leapp.common : parse_leapp_report | Default upgrade_inhibited to false] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:12 ok: [managed-node01] => {"ansible_facts": {"upgrade_inhibited": false}, "changed": false} TASK [infra.leapp.common : parse_leapp_report | Collect human readable report results] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:16 ok: [managed-node01] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false} TASK [infra.leapp.common : parse_leapp_report | Collect JSON report results] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:22 ok: [managed-node01] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false} TASK [infra.leapp.common : parse_leapp_report | Parse report results] ********** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:28 ok: [managed-node01] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false} TASK [infra.leapp.common : parse_leapp_report | Clear leapp_inhibitors] ******** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:35 ok: [managed-node01] => {"ansible_facts": {"leapp_inhibitors": []}, "changed": false} TASK [infra.leapp.common : parse_leapp_report | Check for inhibitors] ********** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:39 ok: [managed-node01] => (item={'audience': 'sysadmin', 'groups': ['error'], 'key': '7ec8269784db1bba2ac54ae438689ef397e16833', 'severity': 'high', 'summary': '{"details": "Command [\'subscription-manager\', \'release\'] failed with exit code 1.", "stderr": "This system is not yet registered. Try \'subscription-manager register --help\' for more information.\\n", "hint": "Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.", "link": "https://access.redhat.com/solutions/6138372"}', 'title': 'A subscription-manager command failed to execute', 'timeStamp': '2026-02-17T13:23:27.285726Z', 'hostname': 'managed-node01', 'actor': 'scan_subscription_manager_info', 'id': '671fc7de204d0bf5b5c69989d9d969c959f9407dc5fc75ca36efd650b128fdc9'}) => {"ansible_facts": {"leapp_inhibitors": [{"actor": "scan_subscription_manager_info", "audience": "sysadmin", "groups": ["error"], "hostname": "managed-node01", "id": "671fc7de204d0bf5b5c69989d9d969c959f9407dc5fc75ca36efd650b128fdc9", "key": "7ec8269784db1bba2ac54ae438689ef397e16833", "severity": "high", "summary": "{\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "timeStamp": "2026-02-17T13:23:27.285726Z", "title": "A subscription-manager command failed to execute"}], "upgrade_inhibited": true}, "ansible_loop_var": "item", "changed": false, "item": {"actor": "scan_subscription_manager_info", "audience": "sysadmin", "groups": ["error"], "hostname": "managed-node01", "id": "671fc7de204d0bf5b5c69989d9d969c959f9407dc5fc75ca36efd650b128fdc9", "key": "7ec8269784db1bba2ac54ae438689ef397e16833", "severity": "high", "summary": "{\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "timeStamp": "2026-02-17T13:23:27.285726Z", "title": "A subscription-manager command failed to execute"}} skipping: [managed-node01] => (item={'audience': 'sysadmin', 'detail': {'remediations': [{'context': 'If some of excluded repositories are still required to be used during the upgrade, execute leapp with the --enablerepo option with the repoid of the repository required to be enabled as an argument (the option can be used multiple times).', 'type': 'hint'}]}, 'groups': ['repository', 'failure'], 'key': '1b9132cb2362ae7830e48eee7811be9527747de8', 'severity': 'info', 'summary': 'The following repositories are not supported by Red Hat and are excluded from the list of repositories used during the upgrade.\n- codeready-builder-for-rhel-10-ppc64le-rpms\n- rhui-codeready-builder-for-rhel-10-aarch64-rhui-rpms\n- codeready-builder-beta-for-rhel-10-x86_64-rpms\n- codeready-builder-beta-for-rhel-10-ppc64le-rpms\n- codeready-builder-for-rhel-10-s390x-eus-rpms\n- codeready-builder-for-rhel-10-aarch64-eus-rpms\n- codeready-builder-for-rhel-10-x86_64-eus-rpms\n- codeready-builder-beta-for-rhel-10-aarch64-rpms\n- codeready-builder-for-rhel-10-ppc64le-eus-rpms\n- codeready-builder-beta-for-rhel-10-s390x-rpms\n- codeready-builder-for-rhel-10-x86_64-rpms\n- rhui-codeready-builder-for-rhel-10-x86_64-rhui-rpms\n- codeready-builder-for-rhel-10-rhui-rpms\n- codeready-builder-for-rhel-10-aarch64-rpms\n- codeready-builder-for-rhel-10-s390x-rpms\n- crb', 'title': 'Excluded target system repositories', 'timeStamp': '2026-02-17T13:23:43.886245Z', 'hostname': 'managed-node01', 'actor': 'repositories_blacklist', 'id': '52d6815cf7a8273cd99cf12431a8ac558f7d9d8a0fa5eda4b6c458988b5df4ed'}) => {"ansible_loop_var": "item", "changed": false, "item": {"actor": "repositories_blacklist", "audience": "sysadmin", "detail": {"remediations": [{"context": "If some of excluded repositories are still required to be used during the upgrade, execute leapp with the --enablerepo option with the repoid of the repository required to be enabled as an argument (the option can be used multiple times).", "type": "hint"}]}, "groups": ["repository", "failure"], "hostname": "managed-node01", "id": "52d6815cf7a8273cd99cf12431a8ac558f7d9d8a0fa5eda4b6c458988b5df4ed", "key": "1b9132cb2362ae7830e48eee7811be9527747de8", "severity": "info", "summary": "The following repositories are not supported by Red Hat and are excluded from the list of repositories used during the upgrade.\n- codeready-builder-for-rhel-10-ppc64le-rpms\n- rhui-codeready-builder-for-rhel-10-aarch64-rhui-rpms\n- codeready-builder-beta-for-rhel-10-x86_64-rpms\n- codeready-builder-beta-for-rhel-10-ppc64le-rpms\n- codeready-builder-for-rhel-10-s390x-eus-rpms\n- codeready-builder-for-rhel-10-aarch64-eus-rpms\n- codeready-builder-for-rhel-10-x86_64-eus-rpms\n- codeready-builder-beta-for-rhel-10-aarch64-rpms\n- codeready-builder-for-rhel-10-ppc64le-eus-rpms\n- codeready-builder-beta-for-rhel-10-s390x-rpms\n- codeready-builder-for-rhel-10-x86_64-rpms\n- rhui-codeready-builder-for-rhel-10-x86_64-rhui-rpms\n- codeready-builder-for-rhel-10-rhui-rpms\n- codeready-builder-for-rhel-10-aarch64-rpms\n- codeready-builder-for-rhel-10-s390x-rpms\n- crb", "timeStamp": "2026-02-17T13:23:43.886245Z", "title": "Excluded target system repositories"}, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : parse_leapp_report | Collect inhibitors] ************ task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:51 ok: [managed-node01] => {"changed": false, "cmd": ["awk", "/\\(inhibitor\\)/,/^-------/", "/var/log/leapp/leapp-report.txt"], "delta": "0:00:00.004040", "end": "2026-02-17 08:24:18.913161", "failed_when_result": false, "msg": "", "rc": 0, "start": "2026-02-17 08:24:18.909121", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [infra.leapp.common : parse_leapp_report | Collect high errors] *********** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/parse_leapp_report.yml:60 ok: [managed-node01] => {"changed": false, "cmd": ["awk", "/high \\(error\\)/,/^-------/", "/var/log/leapp/leapp-report.txt"], "delta": "0:00:00.004036", "end": "2026-02-17 08:24:19.267439", "failed_when_result": false, "msg": "", "rc": 0, "start": "2026-02-17 08:24:19.263403", "stderr": "", "stderr_lines": [], "stdout": "Risk Factor: high (error)\nTitle: A subscription-manager command failed to execute\nSummary: {\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}\nKey: 7ec8269784db1bba2ac54ae438689ef397e16833\n----------------------------------------", "stdout_lines": ["Risk Factor: high (error)", "Title: A subscription-manager command failed to execute", "Summary: {\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "Key: 7ec8269784db1bba2ac54ae438689ef397e16833", "----------------------------------------"]} TASK [infra.leapp.analysis : analysis-leapp | Copy reports to the controller] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:115 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/copy_reports_to_controller.yml for managed-node01 TASK [infra.leapp.analysis : copy_reports_to_controller | Ensure reports directory on controller] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/copy_reports_to_controller.yml:10 changed: [managed-node01 -> localhost] => {"changed": true, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tests/leapp_reports_2026-02-17_13-24-19", "secontext": "unconfined_u:object_r:admin_home_t:s0", "size": 6, "state": "directory", "uid": 0} TASK [infra.leapp.analysis : copy_reports_to_controller | Copy report files to the controller] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/copy_reports_to_controller.yml:18 changed: [managed-node01] => (item={'src': '/var/log/leapp/leapp-report.json', 'ext': 'json'}) => {"ansible_loop_var": "item", "changed": true, "checksum": "a44565fcf71c6bb2d7a2db601181dc61e365fcf7", "dest": "/root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tests/leapp_reports_2026-02-17_13-24-19/leapp-report.managed-node01.json/managed-node01/var/log/leapp/leapp-report.json", "item": {"ext": "json", "src": "/var/log/leapp/leapp-report.json"}, "md5sum": "e9fa11f9a11ac31e0edf41b423968294", "remote_checksum": "a44565fcf71c6bb2d7a2db601181dc61e365fcf7", "remote_md5sum": null} changed: [managed-node01] => (item={'src': '/var/log/leapp/leapp-report.txt', 'ext': 'txt'}) => {"ansible_loop_var": "item", "changed": true, "checksum": "5b8d6f66575fd6bcfdf679dd6d49c929c7412053", "dest": "/root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tests/leapp_reports_2026-02-17_13-24-19/leapp-report.managed-node01.txt/managed-node01/var/log/leapp/leapp-report.txt", "item": {"ext": "txt", "src": "/var/log/leapp/leapp-report.txt"}, "md5sum": "5b431db5ccd4eb8bea933e017c57d3b9", "remote_checksum": "5b8d6f66575fd6bcfdf679dd6d49c929c7412053", "remote_md5sum": null} TASK [infra.leapp.analysis : analysis-leapp | Create remediation hostvars] ***** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:119 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/create_remediation_hostvars.yml for managed-node01 TASK [infra.leapp.analysis : create_remediation_hostvars | Clear remediation_todo] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/create_remediation_hostvars.yml:2 ok: [managed-node01] => {"ansible_facts": {"leapp_remediation_todo": []}, "changed": false} TASK [infra.leapp.analysis : create_remediation_hostvars | Map inhibitors to remediation_todo] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/create_remediation_hostvars.yml:6 failed: [managed-node01] (item=None) => {"ansible_loop_var": "inhibitor", "changed": false, "inhibitor": {"actor": "scan_subscription_manager_info", "audience": "sysadmin", "groups": ["error"], "hostname": "managed-node01", "id": "671fc7de204d0bf5b5c69989d9d969c959f9407dc5fc75ca36efd650b128fdc9", "key": "7ec8269784db1bba2ac54ae438689ef397e16833", "severity": "high", "summary": "{\"details\": \"Command ['subscription-manager', 'release'] failed with exit code 1.\", \"stderr\": \"This system is not yet registered. Try 'subscription-manager register --help' for more information.\\n\", \"hint\": \"Please ensure you have a valid RHEL subscription and your network is up. If you are using proxy for Red Hat subscription-manager, please make sure it is specified inside the /etc/rhsm/rhsm.conf file. Or use the --no-rhsm option when running leapp, if you do not want to use subscription-manager for the in-place upgrade and you want to deliver all target repositories by yourself or using RHUI on public cloud.\", \"link\": \"https://access.redhat.com/solutions/6138372\"}", "timeStamp": "2026-02-17T13:23:27.285726Z", "title": "A subscription-manager command failed to execute"}, "msg": "Failed to template loop_control.label: {{ __leapp_inhibitors_key_map[inhibitor_key]\n if inhibitor_key in __leapp_inhibitors_key_map\n else __leapp_inhibitors_title_map[inhibitor_title] }}: 'dict object' has no attribute 'A subscription-manager command failed to execute'. 'dict object' has no attribute 'A subscription-manager command failed to execute'. {{ __leapp_inhibitors_key_map[inhibitor_key]\n if inhibitor_key in __leapp_inhibitors_key_map\n else __leapp_inhibitors_title_map[inhibitor_title] }}: 'dict object' has no attribute 'A subscription-manager command failed to execute'. 'dict object' has no attribute 'A subscription-manager command failed to execute'", "skip_reason": "Conditional result was False"} TASK [Cleanup | Remove log files] ********************************************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tests/tests_default.yml:13 changed: [managed-node01] => {"changed": true, "cmd": "set -euxo pipefail\nrm -f /var/log/leapp/leapp-upgrade.log\nrm -f /var/log/ripu/ripu.log*\n", "delta": "0:00:01.005900", "end": "2026-02-17 08:24:21.752810", "msg": "", "rc": 0, "start": "2026-02-17 08:24:20.746910", "stderr": "+ rm -f /var/log/leapp/leapp-upgrade.log\n+ rm -f /var/log/ripu/ripu.log", "stderr_lines": ["+ rm -f /var/log/leapp/leapp-upgrade.log", "+ rm -f /var/log/ripu/ripu.log"], "stdout": "", "stdout_lines": []} PLAY RECAP ********************************************************************* managed-node01 : ok=30 changed=10 unreachable=0 failed=1 skipped=10 rescued=0 ignored=0 Feb 17 08:22:58 managed-node01 sshd[4876]: Accepted publickey for root from 10.31.41.208 port 41170 ssh2: ECDSA SHA256:KLW1v6ATZJJzMkxb3g03ESo9SGfLFEvonRaymL9rHI0 Feb 17 08:22:58 managed-node01 systemd-logind[615]: New session 7 of user root. ░░ Subject: A new session 7 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 7 has been created for the user root. ░░ ░░ The leading process of the session is 4876. Feb 17 08:22:58 managed-node01 systemd[1]: Started Session 7 of User root. ░░ Subject: A start job for unit session-7.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-7.scope has finished successfully. ░░ ░░ The job identifier is 1352. Feb 17 08:22:58 managed-node01 sshd[4876]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Feb 17 08:22:59 managed-node01 python3[5053]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Feb 17 08:23:00 managed-node01 python3[5230]: ansible-ansible.builtin.file Invoked with path=/var/log/ripu state=directory owner=root group=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 17 08:23:00 managed-node01 python3[5379]: ansible-ansible.builtin.stat Invoked with path=/var/log/ripu/ripu.log follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 17 08:23:00 managed-node01 python3[5528]: ansible-ansible.legacy.stat Invoked with path=/var/log/ripu/ripu.log follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 17 08:23:01 managed-node01 python3[5648]: ansible-ansible.legacy.copy Invoked with dest=/var/log/ripu/ripu.log owner=root group=root mode=0644 src=/root/.ansible/tmp/ansible-tmp-1771334580.6720881-5680-26692498008197/source _original_basename=tmpew44k2w4 follow=False checksum=d07f0c7b48599f0c4b54e09ada45d9885b771962 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 17 08:23:01 managed-node01 python3[5797]: ansible-ansible.builtin.file Invoked with path=/etc/ansible/facts.d state=directory mode=0755 owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 17 08:23:02 managed-node01 python3[5946]: ansible-ansible.legacy.stat Invoked with path=/etc/ansible/facts.d/pre_ripu.fact follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 17 08:23:02 managed-node01 python3[6066]: ansible-ansible.legacy.copy Invoked with dest=/etc/ansible/facts.d/pre_ripu.fact mode=0644 owner=root group=root src=/root/.ansible/tmp/ansible-tmp-1771334581.8351882-5710-132141815384946/source _original_basename=tmpfegoeqs3 follow=False checksum=1b348fd1227c8d47b88a0a2c38947fb1940bc624 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 17 08:23:02 managed-node01 python3[6215]: ansible-ansible.legacy.command Invoked with _raw_params=set -o pipefail; export PATH=$PATH; rpm -qa | grep -ve '[\.|+]el9' | grep -vE '^(gpg-pubkey|libmodulemd|katello-ca-consumer)' | sort _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 17 08:23:03 managed-node01 python3[6369]: ansible-ansible.legacy.stat Invoked with path=/etc/ansible/facts.d/non_rhel_packages.fact follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 17 08:23:03 managed-node01 python3[6489]: ansible-ansible.legacy.copy Invoked with dest=/etc/ansible/facts.d/non_rhel_packages.fact mode=0644 owner=root group=root src=/root/.ansible/tmp/ansible-tmp-1771334583.1583416-5741-258761498105593/source _original_basename=tmp7msijhej follow=False checksum=97d170e1550eee4afc0af065b78cda302a97674c backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Feb 17 08:23:04 managed-node01 python3[6638]: ansible-ansible.legacy.dnf Invoked with name=['leapp-upgrade'] enablerepo=[] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Feb 17 08:23:10 managed-node01 dbus-broker-launch[607]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Feb 17 08:23:10 managed-node01 dbus-broker-launch[607]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Feb 17 08:23:11 managed-node01 dbus-broker-launch[607]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Feb 17 08:23:11 managed-node01 dbus-broker-launch[607]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Feb 17 08:23:11 managed-node01 dbus-broker-launch[607]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Feb 17 08:23:14 managed-node01 systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-r5fc1a166433a4cc5bd40694525a53fda.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-r5fc1a166433a4cc5bd40694525a53fda.service has finished successfully. ░░ ░░ The job identifier is 1442. Feb 17 08:23:14 managed-node01 systemd[1]: Starting man-db-cache-update.service... ░░ Subject: A start job for unit man-db-cache-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has begun execution. ░░ ░░ The job identifier is 1523. Feb 17 08:23:14 managed-node01 systemd[1]: Reloading. Feb 17 08:23:14 managed-node01 systemd-rc-local-generator[6677]: /etc/rc.d/rc.local is not marked executable, skipping. Feb 17 08:23:14 managed-node01 systemd[1]: Queuing reload/restart jobs for marked units… Feb 17 08:23:15 managed-node01 python3[6928]: ansible-ansible.builtin.file Invoked with path=/var/log/leapp state=directory owner=root group=root mode=0700 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Feb 17 08:23:15 managed-node01 ansible-async_wrapper.py[7347]: Invoked with j108107829421 7200 /root/.ansible/tmp/ansible-tmp-1771334595.4150028-5827-607353698208/AnsiballZ_command.py _ Feb 17 08:23:15 managed-node01 ansible-async_wrapper.py[7350]: Starting module and watcher Feb 17 08:23:15 managed-node01 ansible-async_wrapper.py[7350]: Start watching 7351 (7200) Feb 17 08:23:15 managed-node01 ansible-async_wrapper.py[7351]: Start module (7351) Feb 17 08:23:15 managed-node01 ansible-async_wrapper.py[7347]: Return async_wrapper task started. Feb 17 08:23:16 managed-node01 python3[7352]: ansible-ansible.legacy.command Invoked with executable=/bin/bash _raw_params=set -o pipefail; export PATH=$PATH; ulimit -n 16384; leapp preupgrade --report-schema=1.2.0 2>&1 | tee -a /var/log/ripu/ripu.log _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None creates=None removes=None stdin=None Feb 17 08:23:16 managed-node01 systemd[1]: man-db-cache-update.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit man-db-cache-update.service has successfully entered the 'dead' state. Feb 17 08:23:16 managed-node01 systemd[1]: Finished man-db-cache-update.service. ░░ Subject: A start job for unit man-db-cache-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has finished successfully. ░░ ░░ The job identifier is 1523. Feb 17 08:23:16 managed-node01 systemd[1]: run-r5fc1a166433a4cc5bd40694525a53fda.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-r5fc1a166433a4cc5bd40694525a53fda.service has successfully entered the 'dead' state. Feb 17 08:23:20 managed-node01 ansible-async_wrapper.py[7350]: 7351 still running (7200) Feb 17 08:23:25 managed-node01 ansible-async_wrapper.py[7350]: 7351 still running (7195) Feb 17 08:23:30 managed-node01 ansible-async_wrapper.py[7350]: 7351 still running (7190) Feb 17 08:23:35 managed-node01 ansible-async_wrapper.py[7350]: 7351 still running (7185) Feb 17 08:23:35 managed-node01 systemd[1]: proc-sys-fs-binfmt_misc.automount: Got automount request for /proc/sys/fs/binfmt_misc, triggered by 9788 (sysctl) Feb 17 08:23:35 managed-node01 systemd[1]: Mounting Arbitrary Executable File Formats File System... ░░ Subject: A start job for unit proc-sys-fs-binfmt_misc.mount has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit proc-sys-fs-binfmt_misc.mount has begun execution. ░░ ░░ The job identifier is 1604. Feb 17 08:23:35 managed-node01 systemd[1]: Mounted Arbitrary Executable File Formats File System. ░░ Subject: A start job for unit proc-sys-fs-binfmt_misc.mount has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit proc-sys-fs-binfmt_misc.mount has finished successfully. ░░ ░░ The job identifier is 1604. Feb 17 08:23:40 managed-node01 ansible-async_wrapper.py[7350]: 7351 still running (7180) Feb 17 08:23:45 managed-node01 ansible-async_wrapper.py[7351]: Module complete (7351) Feb 17 08:23:45 managed-node01 ansible-async_wrapper.py[7350]: Done in kid B. Feb 17 08:24:15 managed-node01 sshd[4879]: Received disconnect from 10.31.41.208 port 41170:11: disconnected by user Feb 17 08:24:15 managed-node01 sshd[4879]: Disconnected from user root 10.31.41.208 port 41170 Feb 17 08:24:15 managed-node01 sshd[4876]: pam_unix(sshd:session): session closed for user root Feb 17 08:24:15 managed-node01 systemd[1]: session-7.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-7.scope has successfully entered the 'dead' state. Feb 17 08:24:15 managed-node01 systemd[1]: session-7.scope: Consumed 35.377s CPU time. ░░ Subject: Resources consumed by unit runtime ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-7.scope completed and consumed the indicated resources. Feb 17 08:24:15 managed-node01 systemd-logind[615]: Session 7 logged out. Waiting for processes to exit. Feb 17 08:24:15 managed-node01 systemd-logind[615]: Removed session 7. ░░ Subject: Session 7 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 7 has been terminated. Feb 17 08:24:16 managed-node01 sshd[11167]: Accepted publickey for root from 10.31.41.208 port 50032 ssh2: ECDSA SHA256:KLW1v6ATZJJzMkxb3g03ESo9SGfLFEvonRaymL9rHI0 Feb 17 08:24:16 managed-node01 systemd-logind[615]: New session 8 of user root. ░░ Subject: A new session 8 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 8 has been created for the user root. ░░ ░░ The leading process of the session is 11167. Feb 17 08:24:16 managed-node01 systemd[1]: Started Session 8 of User root. ░░ Subject: A start job for unit session-8.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-8.scope has finished successfully. ░░ ░░ The job identifier is 1608. Feb 17 08:24:16 managed-node01 sshd[11167]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Feb 17 08:24:16 managed-node01 python3[11318]: ansible-ansible.legacy.async_status Invoked with jid=j108107829421.7347 mode=status _async_dir=/root/.ansible_async Feb 17 08:24:16 managed-node01 python3[11414]: ansible-ansible.legacy.async_status Invoked with jid=j108107829421.7347 mode=cleanup _async_dir=/root/.ansible_async Feb 17 08:24:17 managed-node01 python3[11563]: ansible-ansible.builtin.stat Invoked with path=/var/log/leapp/leapp-report.txt follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Feb 17 08:24:18 managed-node01 python3[12012]: ansible-ansible.legacy.command Invoked with _raw_params=awk '/\(inhibitor\)/,/^-------/' /var/log/leapp/leapp-report.txt _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 17 08:24:19 managed-node01 python3[12162]: ansible-ansible.legacy.command Invoked with _raw_params=awk '/high \(error\)/,/^-------/' /var/log/leapp/leapp-report.txt _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Feb 17 08:24:19 managed-node01 python3[12312]: ansible-ansible.legacy.stat Invoked with path=/var/log/leapp/leapp-report.json follow=True get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 17 08:24:20 managed-node01 python3[12486]: ansible-ansible.legacy.stat Invoked with path=/var/log/leapp/leapp-report.txt follow=True get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Feb 17 08:24:20 managed-node01 python3[12660]: ansible-ansible.legacy.command Invoked with executable=/bin/bash _raw_params=set -euxo pipefail rm -f /var/log/leapp/leapp-upgrade.log rm -f /var/log/ripu/ripu.log* _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None creates=None removes=None stdin=None Feb 17 08:24:21 managed-node01 sshd[12687]: Accepted publickey for root from 10.31.41.208 port 50038 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Feb 17 08:24:21 managed-node01 systemd-logind[615]: New session 9 of user root. ░░ Subject: A new session 9 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 9 has been created for the user root. ░░ ░░ The leading process of the session is 12687. Feb 17 08:24:21 managed-node01 systemd[1]: Started Session 9 of User root. ░░ Subject: A start job for unit session-9.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-9.scope has finished successfully. ░░ ░░ The job identifier is 1693. Feb 17 08:24:21 managed-node01 sshd[12687]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Feb 17 08:24:22 managed-node01 sshd[12690]: Received disconnect from 10.31.41.208 port 50038:11: disconnected by user Feb 17 08:24:22 managed-node01 sshd[12690]: Disconnected from user root 10.31.41.208 port 50038 Feb 17 08:24:22 managed-node01 sshd[12687]: pam_unix(sshd:session): session closed for user root Feb 17 08:24:22 managed-node01 systemd[1]: session-9.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-9.scope has successfully entered the 'dead' state. Feb 17 08:24:22 managed-node01 systemd-logind[615]: Session 9 logged out. Waiting for processes to exit. Feb 17 08:24:22 managed-node01 systemd-logind[615]: Removed session 9. ░░ Subject: Session 9 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 9 has been terminated. Feb 17 08:24:22 managed-node01 sshd[12715]: Accepted publickey for root from 10.31.41.208 port 50048 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Feb 17 08:24:22 managed-node01 systemd-logind[615]: New session 10 of user root. ░░ Subject: A new session 10 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 10 has been created for the user root. ░░ ░░ The leading process of the session is 12715. Feb 17 08:24:22 managed-node01 systemd[1]: Started Session 10 of User root. ░░ Subject: A start job for unit session-10.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-10.scope has finished successfully. ░░ ░░ The job identifier is 1778. Feb 17 08:24:22 managed-node01 sshd[12715]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0)