[WARNING]: Collection infra.leapp does not support Ansible version 2.14.18 [WARNING]: running playbook inside collection infra.leapp ansible-playbook [core 2.14.18] config file = /etc/ansible/ansible.cfg configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.9/site-packages/ansible ansible collection location = /root/.ansible/collections:/usr/share/ansible/collections executable location = /usr/bin/ansible-playbook python version = 3.9.25 (main, Mar 9 2026, 00:00:00) [GCC 11.5.0 20240719 (Red Hat 11.5.0-14)] (/usr/bin/python3) jinja version = 3.1.2 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_remediations_9to10.yml ***************************************** 1 plays in /root/.ansible/collections/ansible_collections/infra/leapp/tests/tests_remediations_9to10.yml PLAY [Test RHEL 9 to 10 remediations] ****************************************** TASK [Gathering Facts] ********************************************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tests_remediations_9to10.yml:2 ok: [managed-node02] TASK [Include tests_upgrade_custom playbook] *********************************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tests_remediations_9to10.yml:22 included: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/tests_upgrade_custom.yml for managed-node02 TASK [tests_upgrade_custom | Check if leapp upgrade log exists] **************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/tests_upgrade_custom.yml:12 ok: [managed-node02] => {"changed": false, "stat": {"exists": false}} TASK [tests_upgrade_custom | Skip test if already upgraded or not RHEL {{ rhel_base_ver }}] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/tests_upgrade_custom.yml:17 META: end_play conditional evaluated to False, continuing play skipping: [managed-node02] => {"msg": "end_play", "skip_reason": "end_play conditional evaluated to False, continuing play"} TASK [tests_upgrade_custom | Include common upgrade tasks] ********************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/tests_upgrade_custom.yml:27 included: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/common_upgrade_tasks.yml for managed-node02 TASK [common_upgrade_tasks | Remove leapp packages] **************************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/common_upgrade_tasks.yml:6 ok: [managed-node02] => {"changed": false, "msg": "Nothing to do", "rc": 0, "results": []} TASK [common_upgrade_tasks | Gather setup tasks] ******************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/common_upgrade_tasks.yml:11 ok: [managed-node02 -> localhost] => {"changed": false, "examined": 4, "files": [{"atime": 1773834325.42296, "ctime": 1773834325.2559614, "dev": 51716, "gid": 0, "gr_name": "root", "inode": 746586270, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mode": "0644", "mtime": 1773834325.2559614, "nlink": 1, "path": "/root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/setup/remediate_cifs.yml", "pw_name": "root", "rgrp": true, "roth": true, "rusr": true, "size": 272, "uid": 0, "wgrp": false, "woth": false, "wusr": true, "xgrp": false, "xoth": false, "xusr": false}, {"atime": 1773834325.42296, "ctime": 1773834325.2559614, "dev": 51716, "gid": 0, "gr_name": "root", "inode": 746586271, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mode": "0644", "mtime": 1773834325.2559614, "nlink": 1, "path": "/root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/setup/remediate_remote_using_root.yml", "pw_name": "root", "rgrp": true, "roth": true, "rusr": true, "size": 268, "uid": 0, "wgrp": false, "woth": false, "wusr": true, "xgrp": false, "xoth": false, "xusr": false}, {"atime": 1773834325.42296, "ctime": 1773834325.2559614, "dev": 51716, "gid": 0, "gr_name": "root", "inode": 746586272, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mode": "0644", "mtime": 1773834325.2559614, "nlink": 1, "path": "/root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/setup/remediate_removed_kernel_drivers.yml", "pw_name": "root", "rgrp": true, "roth": true, "rusr": true, "size": 913, "uid": 0, "wgrp": false, "woth": false, "wusr": true, "xgrp": false, "xoth": false, "xusr": false}, {"atime": 1773834325.42296, "ctime": 1773834325.2559614, "dev": 51716, "gid": 0, "gr_name": "root", "inode": 746586273, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mode": "0644", "mtime": 1773834325.2559614, "nlink": 1, "path": "/root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/setup/version_lock.yml", "pw_name": "root", "rgrp": true, "roth": true, "rusr": true, "size": 548, "uid": 0, "wgrp": false, "woth": false, "wusr": true, "xgrp": false, "xoth": false, "xusr": false}], "matched": 4, "msg": "All paths examined", "skipped_paths": {}} TASK [common_upgrade_tasks | Do remediation setup tasks] *********************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/common_upgrade_tasks.yml:20 included: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/setup/remediate_cifs.yml for managed-node02 => (item=/root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/setup/remediate_cifs.yml) included: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/setup/remediate_remote_using_root.yml for managed-node02 => (item=/root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/setup/remediate_remote_using_root.yml) included: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/setup/remediate_removed_kernel_drivers.yml for managed-node02 => (item=/root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/setup/remediate_removed_kernel_drivers.yml) TASK [setup | remediate_cifs | Add a CIFS share to /etc/fstab] ***************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/setup/remediate_cifs.yml:3 changed: [managed-node02] => {"backup": "", "changed": true, "msg": "line added"} TASK [setup | remediate_remote_using_root | Set the parameter to not remediate SSH password authentication] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/setup/remediate_remote_using_root.yml:3 ok: [managed-node02] => {"ansible_facts": {"leapp_remediate_ssh_password_auth": false}, "changed": false} TASK [setup | remediate_removed_kernel_drivers | Set list of test kernel modules] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/setup/remediate_removed_kernel_drivers.yml:4 ok: [managed-node02] => {"ansible_facts": {"leapp_test_kernel_modules": ["qla3xxx", "vmw_pvrdma"]}, "changed": false} TASK [setup | remediate_removed_kernel_drivers | Load the test kernel modules] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/setup/remediate_removed_kernel_drivers.yml:21 TASK [infra.leapp.common : manage_kernel_modules | Load or unload kernel modules] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/manage_kernel_modules.yml:5 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/manage_one_kernel_module.yml for managed-node02 => (item=qla3xxx) included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/manage_one_kernel_module.yml for managed-node02 => (item=vmw_pvrdma) TASK [infra.leapp.common : manage_one_kernel_module | Load or unload kernel module] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/manage_one_kernel_module.yml:5 changed: [managed-node02] => {"changed": true, "name": "qla3xxx", "params": "", "state": "present"} TASK [infra.leapp.common : manage_one_kernel_module | Disable modules-load.d file entry] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/manage_one_kernel_module.yml:17 skipping: [managed-node02] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : manage_one_kernel_module | Ensure modules are not loaded at boot] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/manage_one_kernel_module.yml:25 skipping: [managed-node02] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : manage_one_kernel_module | Debug modprobe.d file] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/manage_one_kernel_module.yml:33 skipping: [managed-node02] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : manage_one_kernel_module | Debug modules-load.d file] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/manage_one_kernel_module.yml:37 skipping: [managed-node02] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : manage_one_kernel_module | Load or unload kernel module] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/manage_one_kernel_module.yml:5 changed: [managed-node02] => {"changed": true, "name": "vmw_pvrdma", "params": "", "state": "present"} TASK [infra.leapp.common : manage_one_kernel_module | Disable modules-load.d file entry] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/manage_one_kernel_module.yml:17 skipping: [managed-node02] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : manage_one_kernel_module | Ensure modules are not loaded at boot] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/manage_one_kernel_module.yml:25 skipping: [managed-node02] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : manage_one_kernel_module | Debug modprobe.d file] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/manage_one_kernel_module.yml:33 skipping: [managed-node02] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : manage_one_kernel_module | Debug modules-load.d file] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/manage_one_kernel_module.yml:37 skipping: [managed-node02] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [common_upgrade_tasks | Do setup tasks] *********************************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/common_upgrade_tasks.yml:31 skipping: [managed-node02] => (item=/root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/setup/version_lock.yml) => {"ansible_loop_var": "setup_task_file", "changed": false, "setup_task_file": "/root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/setup/version_lock.yml", "skip_reason": "Conditional result was False"} skipping: [managed-node02] => {"changed": false, "msg": "All items skipped"} TASK [common_upgrade_tasks | Run first analysis] ******************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/common_upgrade_tasks.yml:42 TASK [infra.leapp.analysis : Lock timestamped variables] *********************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/main.yml:5 ok: [managed-node02] => {"ansible_facts": {"__leapp_timestamp": "2026-03-18_11-51-26"}, "changed": false} TASK [Initialize lock, logging, and common vars] ******************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/main.yml:9 TASK [infra.leapp.common : init_leapp_log | Ensure that log directory exists] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:9 changed: [managed-node02] => {"changed": true, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/var/log/leapp", "secontext": "unconfined_u:object_r:var_log_t:s0", "size": 6, "state": "directory", "uid": 0} TASK [infra.leapp.common : init_leapp_log | Check for existing log file] ******* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:17 ok: [managed-node02] => {"changed": false, "stat": {"exists": false}} TASK [infra.leapp.common : init_leapp_log | Fail if log file already exists] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:22 skipping: [managed-node02] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : init_leapp_log | Create new log file] *************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:31 changed: [managed-node02] => {"changed": true, "checksum": "674a97b9398474d58604975ae0cfe0ddaa34224e", "dest": "/var/log/leapp/ansible_leapp_analysis.log", "gid": 0, "group": "root", "md5sum": "97329bef80c89e3d0c0fc0a188b2e10f", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:var_log_t:s0", "size": 70, "src": "/root/.ansible/tmp/ansible-tmp-1773834687.9702797-9907-208631081033392/source", "state": "file", "uid": 0} TASK [infra.leapp.common : init_leapp_log | /etc/ansible/facts.d directory exists] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:41 changed: [managed-node02] => {"changed": true, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/ansible/facts.d", "secontext": "unconfined_u:object_r:etc_t:s0", "size": 6, "state": "directory", "uid": 0} TASK [infra.leapp.common : init_leapp_log | Capture current ansible_facts for validation after upgrade] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:49 changed: [managed-node02] => (item=/etc/ansible/facts.d/pre_ipu.fact) => {"ansible_loop_var": "item", "changed": true, "checksum": "861613139b9284fd640a6ba02fa33da792bec589", "dest": "/etc/ansible/facts.d/pre_ipu.fact", "gid": 0, "group": "root", "item": "/etc/ansible/facts.d/pre_ipu.fact", "md5sum": "d5c64de514b84145c8a5983c619ca1ea", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 14278, "src": "/root/.ansible/tmp/ansible-tmp-1773834689.3704882-10053-139111807674425/source", "state": "file", "uid": 0} changed: [managed-node02] => (item=/var/log/leapp/ansible_leapp_analysis.log) => {"ansible_loop_var": "item", "changed": true, "checksum": "861613139b9284fd640a6ba02fa33da792bec589", "dest": "/var/log/leapp/ansible_leapp_analysis.log", "gid": 0, "group": "root", "item": "/var/log/leapp/ansible_leapp_analysis.log", "md5sum": "d5c64de514b84145c8a5983c619ca1ea", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:var_log_t:s0", "size": 14278, "src": "/root/.ansible/tmp/ansible-tmp-1773834690.1393416-10053-122672687777611/source", "state": "file", "uid": 0} TASK [infra.leapp.common : init_leapp_log | Capture a list of non-rhel versioned packages] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:60 ok: [managed-node02] => {"changed": false, "cmd": "set -o pipefail; export PATH=$PATH; rpm -qa | grep -ve '[\\.|+]el9' | grep -vE '^(gpg-pubkey|libmodulemd|katello-ca-consumer)' | sort", "delta": "0:00:00.181284", "end": "2026-03-18 07:51:31.434248", "failed_when_result": false, "msg": "non-zero return code", "rc": 1, "start": "2026-03-18 07:51:31.252964", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [infra.leapp.common : init_leapp_log | Create fact with the non-rhel versioned packages list] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:74 ok: [managed-node02] => {"ansible_facts": {"non_rhel_packages": []}, "changed": false} TASK [infra.leapp.common : init_leapp_log | Capture the list of non-rhel versioned packages in a separate fact file] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/init_leapp_log.yml:78 changed: [managed-node02] => (item=/etc/ansible/facts.d/non_rhel_packages.fact) => {"ansible_loop_var": "item", "changed": true, "checksum": "97d170e1550eee4afc0af065b78cda302a97674c", "dest": "/etc/ansible/facts.d/non_rhel_packages.fact", "gid": 0, "group": "root", "item": "/etc/ansible/facts.d/non_rhel_packages.fact", "md5sum": "d751713988987e9331980363e24189ce", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:etc_t:s0", "size": 2, "src": "/root/.ansible/tmp/ansible-tmp-1773834691.5614707-10134-115228402636475/source", "state": "file", "uid": 0} changed: [managed-node02] => (item=/var/log/leapp/ansible_leapp_analysis.log) => {"ansible_loop_var": "item", "changed": true, "checksum": "97d170e1550eee4afc0af065b78cda302a97674c", "dest": "/var/log/leapp/ansible_leapp_analysis.log", "gid": 0, "group": "root", "item": "/var/log/leapp/ansible_leapp_analysis.log", "md5sum": "d751713988987e9331980363e24189ce", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:var_log_t:s0", "size": 2, "src": "/root/.ansible/tmp/ansible-tmp-1773834692.2197683-10134-50637582498366/source", "state": "file", "uid": 0} TASK [infra.leapp.analysis : Include tasks for preupg assistant analysis] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/main.yml:19 skipping: [managed-node02] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.analysis : Include tasks for leapp preupgrade analysis] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/main.yml:23 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml for managed-node02 TASK [analysis-leapp | Include pre_upgrade.yml] ******************************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:4 TASK [infra.leapp.common : pre_upgrade | Register with Satellite activation key] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/pre_upgrade.yml:3 skipping: [managed-node02] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [pre_upgrade | Include custom_local_repos for local_repos_pre_leapp] ****** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/pre_upgrade.yml:10 TASK [infra.leapp.common : custom_local_repos | Remove old /etc/leapp/files/leapp_upgrade_repositories.repo] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/custom_local_repos.yml:2 skipping: [managed-node02] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : custom_local_repos | Validate repo definitions have baseurl or metalink] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/custom_local_repos.yml:9 skipping: [managed-node02] => {"changed": false, "skipped_reason": "No items in the list"} TASK [infra.leapp.common : custom_local_repos | Enable custom upgrade yum repositories] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/custom_local_repos.yml:16 skipping: [managed-node02] => {"changed": false, "skipped_reason": "No items in the list"} TASK [infra.leapp.common : pre_upgrade | Get package version lock entries] ***** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/pre_upgrade.yml:21 ok: [managed-node02] => {"changed": false, "cmd": ["dnf", "versionlock", "list"], "delta": "0:00:00.312483", "end": "2026-03-18 07:51:33.764751", "failed_when_result": false, "msg": "non-zero return code", "rc": 1, "start": "2026-03-18 07:51:33.452268", "stderr": "No such command: versionlock. Please use /usr/bin/dnf --help\nIt could be a DNF plugin command, try: \"dnf install 'dnf-command(versionlock)'\"", "stderr_lines": ["No such command: versionlock. Please use /usr/bin/dnf --help", "It could be a DNF plugin command, try: \"dnf install 'dnf-command(versionlock)'\""], "stdout": "", "stdout_lines": []} TASK [infra.leapp.common : pre_upgrade | Remove all package version locks] ***** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/pre_upgrade.yml:28 skipping: [managed-node02] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : pre_upgrade | Install packages for upgrade from RHEL 7] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/pre_upgrade.yml:36 changed: [managed-node02] => {"changed": true, "msg": "", "rc": 0, "results": ["Installed: leapp-0.20.0-1.el9.noarch", "Installed: leapp-deps-0.20.0-1.el9.noarch", "Installed: leapp-upgrade-el9toel10-0.23.0-1.el9.noarch", "Installed: leapp-upgrade-el9toel10-deps-0.23.0-1.el9.noarch", "Installed: python3-leapp-0.20.0-1.el9.noarch", "Installed: python3-pyudev-0.22.0-6.el9.noarch", "Installed: systemd-container-252-55.el9_7.7.x86_64", "Installed: policycoreutils-python-utils-3.6-3.el9.noarch", "Installed: pciutils-3.7.0-7.el9.x86_64", "Installed: libdb-utils-5.3.28-57.el9_6.x86_64"]} TASK [infra.leapp.common : pre_upgrade | Include update-and-reboot.yml] ******** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/pre_upgrade.yml:45 fatal: [managed-node02]: FAILED! => {"msg": "The conditional check 'leapp_pre_upgrade_update | bool' failed. The error was: error while evaluating conditional (leapp_pre_upgrade_update | bool): 'leapp_pre_upgrade_update' is undefined. 'leapp_pre_upgrade_update' is undefined\n\nThe error appears to be in '/root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/pre_upgrade.yml': line 45, column 7, but may\nbe elsewhere in the file depending on the exact syntax problem.\n\nThe offending line appears to be:\n\n\n - name: pre_upgrade | Include update-and-reboot.yml\n ^ here\n"} TASK [analysis-leapp | Include custom_local_repos for local_repos_post_analysis] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:70 TASK [infra.leapp.common : custom_local_repos | Remove old /etc/leapp/files/leapp_upgrade_repositories.repo] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/custom_local_repos.yml:2 skipping: [managed-node02] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : custom_local_repos | Validate repo definitions have baseurl or metalink] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/custom_local_repos.yml:9 skipping: [managed-node02] => {"changed": false, "skipped_reason": "No items in the list"} TASK [infra.leapp.common : custom_local_repos | Enable custom upgrade yum repositories] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/custom_local_repos.yml:16 skipping: [managed-node02] => {"changed": false, "skipped_reason": "No items in the list"} TASK [analysis-leapp | Restore original Satellite activation key] ************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:80 skipping: [managed-node02] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [analysis-leapp | Copy reports to the controller] ************************* task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/analysis/tasks/analysis-leapp.yml:91 TASK [infra.leapp.common : copy_reports_to_controller | Ensure reports directory on controller] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_reports_to_controller.yml:20 changed: [managed-node02 -> localhost] => {"changed": true, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/root/.ansible/collections/ansible_collections/infra/leapp/tests/ansible_leapp_analysis_logs_2026-03-18_11-51-26", "secontext": "unconfined_u:object_r:admin_home_t:s0", "size": 6, "state": "directory", "uid": 0} TASK [infra.leapp.common : copy_reports_to_controller | Fetch report files if they exist] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_reports_to_controller.yml:30 included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/fetch_file_if_exists.yml for managed-node02 => (item=/var/log/leapp/leapp-report.txt) included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/fetch_file_if_exists.yml for managed-node02 => (item=/var/log/leapp/leapp-report.json) included: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/fetch_file_if_exists.yml for managed-node02 => (item=/var/log/leapp/leapp-preupgrade.log) TASK [infra.leapp.common : fetch_file_if_exists | Check if file exists] ******** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/fetch_file_if_exists.yml:7 ok: [managed-node02] => {"changed": false, "stat": {"exists": false}} TASK [infra.leapp.common : fetch_file_if_exists | Copy report file to the controller] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/fetch_file_if_exists.yml:12 skipping: [managed-node02] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : fetch_file_if_exists | Check if file exists] ******** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/fetch_file_if_exists.yml:7 ok: [managed-node02] => {"changed": false, "stat": {"exists": false}} TASK [infra.leapp.common : fetch_file_if_exists | Copy report file to the controller] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/fetch_file_if_exists.yml:12 skipping: [managed-node02] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [infra.leapp.common : fetch_file_if_exists | Check if file exists] ******** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/fetch_file_if_exists.yml:7 ok: [managed-node02] => {"changed": false, "stat": {"exists": false}} TASK [infra.leapp.common : fetch_file_if_exists | Copy report file to the controller] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/fetch_file_if_exists.yml:12 skipping: [managed-node02] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [copy_reports_to_controller | Copy log file to the controller] ************ task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_reports_to_controller.yml:39 TASK [infra.leapp.common : copy_archive_leapp_log | Check for log file] ******** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_archive_leapp_log.yml:11 ok: [managed-node02] => {"changed": false, "stat": {"atime": 1773834692.8035798, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "97d170e1550eee4afc0af065b78cda302a97674c", "ctime": 1773834692.8045797, "dev": 51716, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 167772295, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "application/json", "mode": "0644", "mtime": 1773834692.5385776, "nlink": 1, "path": "/var/log/leapp/ansible_leapp_analysis.log", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 2, "uid": 0, "version": "2902122334", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false}} TASK [infra.leapp.common : copy_archive_leapp_log | Add end time to log file] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_archive_leapp_log.yml:19 changed: [managed-node02] => {"backup": "", "changed": true, "msg": "line added"} TASK [infra.leapp.common : copy_archive_leapp_log | Slurp file /var/log/leapp/ansible_leapp_analysis.log] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_archive_leapp_log.yml:27 ok: [managed-node02] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false} TASK [infra.leapp.common : copy_archive_leapp_log | Decode file /var/log/leapp/ansible_leapp_analysis.log] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_archive_leapp_log.yml:33 ok: [managed-node02] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": false} TASK [infra.leapp.common : copy_archive_leapp_log | Ensure reports directory on controller] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_archive_leapp_log.yml:38 ok: [managed-node02 -> localhost] => {"changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/root/.ansible/collections/ansible_collections/infra/leapp/tests/ansible_leapp_analysis_logs_2026-03-18_11-51-26", "secontext": "unconfined_u:object_r:admin_home_t:s0", "size": 6, "state": "directory", "uid": 0} TASK [infra.leapp.common : copy_archive_leapp_log | Copy ansible leapp log to the controller] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_archive_leapp_log.yml:48 changed: [managed-node02] => {"changed": true, "checksum": "6f367f106f3316381b606e9ad52ae26af401d75a", "dest": "/root/.ansible/collections/ansible_collections/infra/leapp/tests/ansible_leapp_analysis_logs_2026-03-18_11-51-26/managed-node02/ansible_leapp_analysis.log", "md5sum": "af88485396ffe318e8d7faf3f49fbd7a", "remote_checksum": "6f367f106f3316381b606e9ad52ae26af401d75a", "remote_md5sum": null} TASK [infra.leapp.common : copy_archive_leapp_log | Copy log file to timestamped location] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_archive_leapp_log.yml:54 changed: [managed-node02] => {"changed": true, "checksum": "6f367f106f3316381b606e9ad52ae26af401d75a", "dest": "/var/log/leapp/ansible_leapp_analysis_2026-03-18_11-51-26.log", "gid": 0, "group": "root", "md5sum": "af88485396ffe318e8d7faf3f49fbd7a", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:var_log_t:s0", "size": 37, "src": "/var/log/leapp/ansible_leapp_analysis.log", "state": "file", "uid": 0} TASK [infra.leapp.common : copy_archive_leapp_log | Remove original log file] *** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/copy_archive_leapp_log.yml:66 changed: [managed-node02] => {"changed": true, "path": "/var/log/leapp/ansible_leapp_analysis.log", "state": "absent"} TASK [tests_upgrade_custom | Include cleanup logs] ***************************** task path: /root/.ansible/collections/ansible_collections/infra/leapp/tests/tasks/tests_upgrade_custom.yml:49 TASK [infra.leapp.common : cleanup_logs | Cleanup | Remove log files] ********** task path: /root/.ansible/collections/ansible_collections/infra/leapp/roles/common/tasks/cleanup_logs.yml:2 changed: [managed-node02] => {"changed": true, "cmd": "set -euxo pipefail\nrm -f /var/log/leapp/*.log\nrm -f /var/log/leapp/*.json\nrm -f /var/log/leapp/*.txt\n", "delta": "0:00:00.005407", "end": "2026-03-18 07:51:44.262961", "msg": "", "rc": 0, "start": "2026-03-18 07:51:44.257554", "stderr": "+ rm -f /var/log/leapp/ansible_leapp_analysis_2026-03-18_11-51-26.log\n+ rm -f '/var/log/leapp/*.json'\n+ rm -f '/var/log/leapp/*.txt'", "stderr_lines": ["+ rm -f /var/log/leapp/ansible_leapp_analysis_2026-03-18_11-51-26.log", "+ rm -f '/var/log/leapp/*.json'", "+ rm -f '/var/log/leapp/*.txt'"], "stdout": "", "stdout_lines": []} PLAY RECAP ********************************************************************* managed-node02 : ok=44 changed=15 unreachable=0 failed=1 skipped=23 rescued=0 ignored=0 Mar 18 07:51:20 managed-node02 python3[7969]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Mar 18 07:51:22 managed-node02 python3[8145]: ansible-ansible.builtin.stat Invoked with path=/var/log/leapp/leapp-upgrade.log follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Mar 18 07:51:22 managed-node02 python3[8294]: ansible-ansible.legacy.dnf Invoked with name=['leapp-upgrade'] state=absent allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Mar 18 07:51:24 managed-node02 python3[8444]: ansible-ansible.builtin.lineinfile Invoked with path=/etc/fstab line=//127.0.0.1/test_remediate_cifs /mnt/cifs cifs username=test,password=test 0 0 state=present backrefs=False create=False backup=False firstmatch=False unsafe_writes=False regexp=None search_string=None insertafter=None insertbefore=None validate=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Mar 18 07:51:25 managed-node02 python3[8593]: ansible-infra.leapp.modprobe Invoked with name=qla3xxx state=present params= persistent=disabled Mar 18 07:51:25 managed-node02 kernel: Warning: Unmaintained driver is detected: qla3xxx Mar 18 07:51:25 managed-node02 kernel: Warning: Unmaintained driver is detected: ql3xxx_driver_init Mar 18 07:51:26 managed-node02 python3[8745]: ansible-infra.leapp.modprobe Invoked with name=vmw_pvrdma state=present params= persistent=disabled Mar 18 07:51:26 managed-node02 kernel: Warning: Unmaintained driver is detected: vmw_pvrdma Mar 18 07:51:26 managed-node02 kernel: Warning: Unmaintained driver is detected: pvrdma_init Mar 18 07:51:27 managed-node02 python3[8902]: ansible-ansible.builtin.file Invoked with path=/var/log/leapp state=directory owner=root group=root mode=0755 recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Mar 18 07:51:27 managed-node02 python3[9051]: ansible-ansible.builtin.stat Invoked with path=/var/log/leapp/ansible_leapp_analysis.log follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Mar 18 07:51:28 managed-node02 python3[9200]: ansible-ansible.legacy.stat Invoked with path=/var/log/leapp/ansible_leapp_analysis.log follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Mar 18 07:51:28 managed-node02 python3[9320]: ansible-ansible.legacy.copy Invoked with dest=/var/log/leapp/ansible_leapp_analysis.log owner=root group=root mode=0644 src=/root/.ansible/tmp/ansible-tmp-1773834687.9702797-9907-208631081033392/source _original_basename=tmpo7myepbm follow=False checksum=674a97b9398474d58604975ae0cfe0ddaa34224e backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Mar 18 07:51:29 managed-node02 python3[9469]: ansible-ansible.builtin.file Invoked with path=/etc/ansible/facts.d state=directory mode=0755 owner=root group=root recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None seuser=None serole=None selevel=None setype=None attributes=None Mar 18 07:51:29 managed-node02 python3[9618]: ansible-ansible.legacy.stat Invoked with path=/etc/ansible/facts.d/pre_ipu.fact follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Mar 18 07:51:30 managed-node02 python3[9738]: ansible-ansible.legacy.copy Invoked with dest=/etc/ansible/facts.d/pre_ipu.fact mode=0644 owner=root group=root src=/root/.ansible/tmp/ansible-tmp-1773834689.3704882-10053-139111807674425/source _original_basename=tmp9gzqf4y3 follow=False checksum=861613139b9284fd640a6ba02fa33da792bec589 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Mar 18 07:51:30 managed-node02 python3[9887]: ansible-ansible.legacy.stat Invoked with path=/var/log/leapp/ansible_leapp_analysis.log follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Mar 18 07:51:30 managed-node02 python3[10009]: ansible-ansible.legacy.copy Invoked with dest=/var/log/leapp/ansible_leapp_analysis.log mode=0644 owner=root group=root src=/root/.ansible/tmp/ansible-tmp-1773834690.1393416-10053-122672687777611/source _original_basename=tmpp718e64b follow=False checksum=861613139b9284fd640a6ba02fa33da792bec589 backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Mar 18 07:51:31 managed-node02 python3[10158]: ansible-ansible.legacy.command Invoked with _raw_params=set -o pipefail; export PATH=$PATH; rpm -qa | grep -ve '[\.|+]el9' | grep -vE '^(gpg-pubkey|libmodulemd|katello-ca-consumer)' | sort _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Mar 18 07:51:31 managed-node02 python3[10312]: ansible-ansible.legacy.stat Invoked with path=/etc/ansible/facts.d/non_rhel_packages.fact follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Mar 18 07:51:32 managed-node02 python3[10432]: ansible-ansible.legacy.copy Invoked with dest=/etc/ansible/facts.d/non_rhel_packages.fact mode=0644 owner=root group=root src=/root/.ansible/tmp/ansible-tmp-1773834691.5614707-10134-115228402636475/source _original_basename=tmphs94khq6 follow=False checksum=97d170e1550eee4afc0af065b78cda302a97674c backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Mar 18 07:51:32 managed-node02 python3[10581]: ansible-ansible.legacy.stat Invoked with path=/var/log/leapp/ansible_leapp_analysis.log follow=False get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Mar 18 07:51:32 managed-node02 python3[10703]: ansible-ansible.legacy.copy Invoked with dest=/var/log/leapp/ansible_leapp_analysis.log mode=0644 owner=root group=root src=/root/.ansible/tmp/ansible-tmp-1773834692.2197683-10134-50637582498366/source _original_basename=tmposfaktjd follow=False checksum=97d170e1550eee4afc0af065b78cda302a97674c backup=False force=True unsafe_writes=False content=NOT_LOGGING_PARAMETER validate=None directory_mode=None remote_src=None local_follow=None seuser=None serole=None selevel=None setype=None attributes=None Mar 18 07:51:33 managed-node02 python3[10852]: ansible-ansible.legacy.command Invoked with _raw_params=dnf versionlock list _uses_shell=False stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Mar 18 07:51:34 managed-node02 python3[11002]: ansible-ansible.legacy.dnf Invoked with name=['leapp-upgrade'] state=latest allow_downgrade=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 allowerasing=False nobest=False conf_file=None disable_excludes=None download_dir=None list=None releasever=None Mar 18 07:51:36 managed-node02 dbus-broker-launch[602]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Mar 18 07:51:36 managed-node02 dbus-broker-launch[602]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Mar 18 07:51:36 managed-node02 dbus-broker-launch[602]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Mar 18 07:51:36 managed-node02 dbus-broker-launch[602]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Mar 18 07:51:36 managed-node02 dbus-broker-launch[602]: Noticed file-system modification, trigger reload. ░░ Subject: A configuration directory was written to ░░ Defined-By: dbus-broker ░░ Support: https://groups.google.com/forum/#!forum/bus1-devel ░░ ░░ A write was detected to one of the directories containing D-Bus configuration ░░ files, triggering a configuration reload. ░░ ░░ This functionality exists for backwards compatibility to pick up changes to ░░ D-Bus configuration without an explicit reolad request. Typically when ░░ installing or removing third-party software causes D-Bus configuration files ░░ to be added or removed. ░░ ░░ It is worth noting that this may cause partial configuration to be loaded in ░░ case dispatching this notification races with the writing of the configuration ░░ files. However, a future notification will then cause the configuration to be ░░ reladed again. Mar 18 07:51:38 managed-node02 systemd[1]: Started /usr/bin/systemctl start man-db-cache-update. ░░ Subject: A start job for unit run-r1f4eb7746c86401688b91b4a16e7c3ba.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit run-r1f4eb7746c86401688b91b4a16e7c3ba.service has finished successfully. ░░ ░░ The job identifier is 1576. Mar 18 07:51:38 managed-node02 systemd[1]: Starting man-db-cache-update.service... ░░ Subject: A start job for unit man-db-cache-update.service has begun execution ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has begun execution. ░░ ░░ The job identifier is 1655. Mar 18 07:51:38 managed-node02 systemd[1]: Reloading. Mar 18 07:51:38 managed-node02 systemd-rc-local-generator[11032]: /etc/rc.d/rc.local is not marked executable, skipping. Mar 18 07:51:38 managed-node02 systemd[1]: Queuing reload/restart jobs for marked units… Mar 18 07:51:39 managed-node02 systemd[1]: man-db-cache-update.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit man-db-cache-update.service has successfully entered the 'dead' state. Mar 18 07:51:39 managed-node02 systemd[1]: Finished man-db-cache-update.service. ░░ Subject: A start job for unit man-db-cache-update.service has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit man-db-cache-update.service has finished successfully. ░░ ░░ The job identifier is 1655. Mar 18 07:51:39 managed-node02 systemd[1]: run-r1f4eb7746c86401688b91b4a16e7c3ba.service: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit run-r1f4eb7746c86401688b91b4a16e7c3ba.service has successfully entered the 'dead' state. Mar 18 07:51:40 managed-node02 python3[11600]: ansible-ansible.builtin.stat Invoked with path=/var/log/leapp/leapp-report.txt follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Mar 18 07:51:40 managed-node02 python3[11749]: ansible-ansible.builtin.stat Invoked with path=/var/log/leapp/leapp-report.json follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Mar 18 07:51:41 managed-node02 python3[11898]: ansible-ansible.builtin.stat Invoked with path=/var/log/leapp/leapp-preupgrade.log follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Mar 18 07:51:41 managed-node02 python3[12047]: ansible-ansible.builtin.stat Invoked with path=/var/log/leapp/ansible_leapp_analysis.log follow=False get_md5=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Mar 18 07:51:41 managed-node02 python3[12198]: ansible-ansible.builtin.lineinfile Invoked with path=/var/log/leapp/ansible_leapp_analysis.log line=Job ended at 2026-03-18T11:51:41Z owner=root group=root mode=0644 state=present backrefs=False create=False backup=False firstmatch=False unsafe_writes=False regexp=None search_string=None insertafter=None insertbefore=None validate=None seuser=None serole=None selevel=None setype=None attributes=None Mar 18 07:51:42 managed-node02 python3[12496]: ansible-ansible.legacy.stat Invoked with path=/var/log/leapp/ansible_leapp_analysis.log follow=True get_checksum=True checksum_algorithm=sha1 get_md5=False get_mime=True get_attributes=True Mar 18 07:51:43 managed-node02 python3[12670]: ansible-ansible.legacy.copy Invoked with src=/var/log/leapp/ansible_leapp_analysis.log dest=/var/log/leapp/ansible_leapp_analysis_2026-03-18_11-51-26.log remote_src=True mode=preserve backup=False force=True follow=False unsafe_writes=False _original_basename=None content=NOT_LOGGING_PARAMETER validate=None directory_mode=None local_follow=None checksum=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Mar 18 07:51:43 managed-node02 python3[12819]: ansible-ansible.builtin.file Invoked with path=/var/log/leapp/ansible_leapp_analysis.log state=absent recurse=False force=False follow=True modification_time_format=%Y%m%d%H%M.%S access_time_format=%Y%m%d%H%M.%S unsafe_writes=False _original_basename=None _diff_peek=None src=None modification_time=None access_time=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Mar 18 07:51:44 managed-node02 python3[12968]: ansible-ansible.legacy.command Invoked with _raw_params=set -euxo pipefail rm -f /var/log/leapp/*.log rm -f /var/log/leapp/*.json rm -f /var/log/leapp/*.txt _uses_shell=True stdin_add_newline=True strip_empty_ends=True argv=None chdir=None executable=None creates=None removes=None stdin=None Mar 18 07:51:44 managed-node02 sshd[12996]: Accepted publickey for root from 10.31.15.33 port 50988 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Mar 18 07:51:44 managed-node02 systemd-logind[609]: New session 11 of user root. ░░ Subject: A new session 11 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 11 has been created for the user root. ░░ ░░ The leading process of the session is 12996. Mar 18 07:51:44 managed-node02 systemd[1]: Started Session 11 of User root. ░░ Subject: A start job for unit session-11.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-11.scope has finished successfully. ░░ ░░ The job identifier is 1734. Mar 18 07:51:44 managed-node02 sshd[12996]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Mar 18 07:51:44 managed-node02 sshd[12999]: Received disconnect from 10.31.15.33 port 50988:11: disconnected by user Mar 18 07:51:44 managed-node02 sshd[12999]: Disconnected from user root 10.31.15.33 port 50988 Mar 18 07:51:44 managed-node02 sshd[12996]: pam_unix(sshd:session): session closed for user root Mar 18 07:51:44 managed-node02 systemd-logind[609]: Session 11 logged out. Waiting for processes to exit. Mar 18 07:51:44 managed-node02 systemd[1]: session-11.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-11.scope has successfully entered the 'dead' state. Mar 18 07:51:44 managed-node02 systemd-logind[609]: Removed session 11. ░░ Subject: Session 11 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 11 has been terminated. Mar 18 07:51:44 managed-node02 sshd[13024]: Accepted publickey for root from 10.31.15.33 port 50998 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Mar 18 07:51:44 managed-node02 systemd-logind[609]: New session 12 of user root. ░░ Subject: A new session 12 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 12 has been created for the user root. ░░ ░░ The leading process of the session is 13024. Mar 18 07:51:44 managed-node02 systemd[1]: Started Session 12 of User root. ░░ Subject: A start job for unit session-12.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-12.scope has finished successfully. ░░ ░░ The job identifier is 1817. Mar 18 07:51:44 managed-node02 sshd[13024]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0)