ansible-playbook 2.9.27 config file = /etc/ansible/ansible.cfg configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python2.7/site-packages/ansible executable location = /usr/bin/ansible-playbook python version = 2.7.5 (default, Nov 14 2023, 16:14:06) [GCC 4.8.5 20150623 (Red Hat 4.8.5-44)] Using /etc/ansible/ansible.cfg as config file [WARNING]: running playbook inside collection fedora.linux_system_roles statically imported: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml statically imported: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml statically imported: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/check_candlepin.yml statically imported: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml statically imported: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml Skipping callback 'actionable', as we already have a stdout callback. Skipping callback 'counter_enabled', as we already have a stdout callback. Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'full_skip', as we already have a stdout callback. Skipping callback 'json', as we already have a stdout callback. Skipping callback 'jsonl', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'null', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. Skipping callback 'selective', as we already have a stdout callback. Skipping callback 'skippy', as we already have a stdout callback. Skipping callback 'stderr', as we already have a stdout callback. Skipping callback 'unixy', as we already have a stdout callback. Skipping callback 'yaml', as we already have a stdout callback. PLAYBOOK: tests_proxy.yml ****************************************************** 1 plays in /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tests_proxy.yml PLAY [Basic proxy test] ******************************************************** TASK [Gathering Facts] ********************************************************* task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tests_proxy.yml:3 Saturday 21 March 2026 15:32:55 -0400 (0:00:00.032) 0:00:00.032 ******** ok: [managed-node1] META: ran handlers TASK [Get LSR_RHC_TEST_DATA environment variable] ****************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml:3 Saturday 21 March 2026 15:32:56 -0400 (0:00:00.816) 0:00:00.848 ******** ok: [managed-node1] => { "ansible_facts": { "lsr_rhc_test_data_file": "" }, "changed": false } TASK [Import test data] ******************************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml:12 Saturday 21 March 2026 15:32:56 -0400 (0:00:00.046) 0:00:00.895 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get facts for external test data] **************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml:16 Saturday 21 March 2026 15:32:56 -0400 (0:00:00.036) 0:00:00.931 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set local lsr_rhc_test_data] ********************************************* task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml:24 Saturday 21 March 2026 15:32:56 -0400 (0:00:00.036) 0:00:00.968 ******** ok: [managed-node1] => { "ansible_facts": { "lsr_rhc_test_data": { "baseurl": "http://localhost:8080", "candlepin_host": "candlepin.local", "candlepin_insecure": false, "candlepin_port": 8443, "candlepin_prefix": "/candlepin", "env_nonworking": "Ceci n'est pas une environment", "envs_register": [ "Environment 2" ], "insights": false, "proxy_auth_hostname": "localhost", "proxy_auth_password": "proxypass", "proxy_auth_port": 3130, "proxy_auth_scheme": "https", "proxy_auth_username": "proxyuser", "proxy_noauth_hostname": "localhost", "proxy_noauth_port": 3128, "proxy_noauth_scheme": "https", "proxy_nonworking_hostname": "wrongproxy", "proxy_nonworking_password": "wrong-proxypassword", "proxy_nonworking_port": 4000, "proxy_nonworking_username": "wrong-proxyuser", "reg_activation_keys": [ "default_key" ], "reg_invalid_password": "invalid-password", "reg_invalid_username": "invalid-user", "reg_organization": "donaldduck", "reg_password": "password", "reg_username": "doc", "release": null, "repositories": [ { "name": "donaldy-content-label-7051", "state": "enabled" }, { "name": "content-label-32060", "state": "disabled" } ] } }, "ansible_included_var_files": [ "/tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/../files/candlepin_data.yml" ], "changed": false } TASK [Check if system is ostree] *********************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml:32 Saturday 21 March 2026 15:32:56 -0400 (0:00:00.042) 0:00:01.010 ******** ok: [managed-node1] => { "changed": false, "stat": { "exists": false } } TASK [Set flag to indicate system is ostree] *********************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml:37 Saturday 21 March 2026 15:32:57 -0400 (0:00:00.411) 0:00:01.422 ******** ok: [managed-node1] => { "ansible_facts": { "__rhc_is_ostree": false }, "changed": false } TASK [Set flag to indicate use of external proxy] ****************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml:41 Saturday 21 March 2026 15:32:57 -0400 (0:00:00.041) 0:00:01.463 ******** ok: [managed-node1] => { "ansible_facts": { "__rhc_external_proxy_url": "", "__rhc_use_external_proxy": false }, "changed": false } TASK [Set rhc_external_proxy to empty when external proxy is not used] ********* task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml:46 Saturday 21 March 2026 15:32:57 -0400 (0:00:00.043) 0:00:01.507 ******** ok: [managed-node1] => { "ansible_facts": { "rhc_external_proxy": {} }, "changed": false } TASK [Parse and set rhc_external_proxy] **************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml:54 Saturday 21 March 2026 15:32:57 -0400 (0:00:00.038) 0:00:01.545 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Add proxy vars to bashrc] ************************************************ task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml:63 Saturday 21 March 2026 15:32:57 -0400 (0:00:00.012) 0:00:01.558 ******** skipping: [managed-node1] => (item=export http_proxy=) => { "ansible_loop_var": "item", "changed": false, "item": "export http_proxy=", "skip_reason": "Conditional result was False" } skipping: [managed-node1] => (item=export https_proxy=) => { "ansible_loop_var": "item", "changed": false, "item": "export https_proxy=", "skip_reason": "Conditional result was False" } TASK [Get facts for external test data] **************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:9 Saturday 21 March 2026 15:32:57 -0400 (0:00:00.041) 0:00:01.599 ******** ok: [managed-node1] TASK [Set helper fact for Candlepin base URL] ********************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:17 Saturday 21 March 2026 15:32:57 -0400 (0:00:00.347) 0:00:01.947 ******** ok: [managed-node1] => { "ansible_facts": { "_cp_url": "https://candlepin.local:8443/candlepin" }, "changed": false } TASK [Set helper fact for Candlepin owner URL] ********************************* task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:21 Saturday 21 March 2026 15:32:57 -0400 (0:00:00.041) 0:00:01.988 ******** ok: [managed-node1] => { "ansible_facts": { "_cp_url_owner": "https://candlepin.local:8443/candlepin/owners/donaldduck" }, "changed": false } TASK [Add candlepin hostname to /etc/hosts] ************************************ task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:25 Saturday 21 March 2026 15:32:57 -0400 (0:00:00.042) 0:00:02.031 ******** changed: [managed-node1] => { "backup": "", "changed": true } MSG: line added TASK [Install needed packages] ************************************************* task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:31 Saturday 21 March 2026 15:32:58 -0400 (0:00:00.379) 0:00:02.410 ******** changed: [managed-node1] => { "changed": true, "changes": { "installed": [ "podman" ] }, "rc": 0, "results": [ "Loaded plugins: fastestmirror, product-id, search-disabled-repos, subscription-\n : manager\n\nThis system is not registered with an entitlement server. You can use subscription-manager to register.\n\nLoading mirror speeds from cached hostfile\nResolving Dependencies\n--> Running transaction check\n---> Package podman.x86_64 0:1.6.4-36.el7_9 will be installed\n--> Processing Dependency: slirp4netns >= 0.4.0-1 for package: podman-1.6.4-36.el7_9.x86_64\n--> Processing Dependency: runc >= 1.0.0-57 for package: podman-1.6.4-36.el7_9.x86_64\n--> Processing Dependency: containers-common >= 0.1.29-3 for package: podman-1.6.4-36.el7_9.x86_64\n--> Processing Dependency: containernetworking-plugins >= 0.8.1-1 for package: podman-1.6.4-36.el7_9.x86_64\n--> Processing Dependency: nftables for package: podman-1.6.4-36.el7_9.x86_64\n--> Processing Dependency: fuse-overlayfs for package: podman-1.6.4-36.el7_9.x86_64\n--> Processing Dependency: container-selinux for package: podman-1.6.4-36.el7_9.x86_64\n--> Processing Dependency: conmon for package: podman-1.6.4-36.el7_9.x86_64\n--> Running transaction check\n---> Package conmon.x86_64 2:2.0.8-1.el7 will be installed\n---> Package container-selinux.noarch 2:2.119.2-1.911c772.el7_8 will be installed\n---> Package containernetworking-plugins.x86_64 0:0.8.3-3.el7.centos will be installed\n---> Package containers-common.x86_64 1:0.1.40-11.el7_8 will be installed\n---> Package fuse-overlayfs.x86_64 0:0.7.2-6.el7_8 will be installed\n--> Processing Dependency: libfuse3.so.3(FUSE_3.2)(64bit) for package: fuse-overlayfs-0.7.2-6.el7_8.x86_64\n--> Processing Dependency: libfuse3.so.3(FUSE_3.0)(64bit) for package: fuse-overlayfs-0.7.2-6.el7_8.x86_64\n--> Processing Dependency: libfuse3.so.3()(64bit) for package: fuse-overlayfs-0.7.2-6.el7_8.x86_64\n---> Package nftables.x86_64 1:0.8-14.el7 will be installed\n--> Processing Dependency: libnftnl.so.7(LIBNFTNL_5)(64bit) for package: 1:nftables-0.8-14.el7.x86_64\n--> Processing Dependency: libnftnl.so.7()(64bit) for package: 1:nftables-0.8-14.el7.x86_64\n---> Package runc.x86_64 0:1.0.0-70.rc10.el7_9 will be installed\n--> Processing Dependency: criu for package: runc-1.0.0-70.rc10.el7_9.x86_64\n---> Package slirp4netns.x86_64 0:0.4.3-4.el7_8 will be installed\n--> Running transaction check\n---> Package criu.x86_64 0:3.12-2.el7 will be installed\n--> Processing Dependency: libprotobuf-c.so.1(LIBPROTOBUF_C_1.0.0)(64bit) for package: criu-3.12-2.el7.x86_64\n--> Processing Dependency: libprotobuf-c.so.1()(64bit) for package: criu-3.12-2.el7.x86_64\n--> Processing Dependency: libnet.so.1()(64bit) for package: criu-3.12-2.el7.x86_64\n---> Package fuse3-libs.x86_64 0:3.6.1-4.el7 will be installed\n---> Package libnftnl.x86_64 0:1.0.8-3.el7 will be installed\n--> Running transaction check\n---> Package libnet.x86_64 0:1.1.6-7.el7 will be installed\n---> Package protobuf-c.x86_64 0:1.0.2-3.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package Arch Version Repository\n Size\n================================================================================\nInstalling:\n podman x86_64 1.6.4-36.el7_9 extras 13 M\nInstalling for dependencies:\n conmon x86_64 2:2.0.8-1.el7 extras 31 k\n container-selinux noarch 2:2.119.2-1.911c772.el7_8 extras 40 k\n containernetworking-plugins x86_64 0.8.3-3.el7.centos extras 20 M\n containers-common x86_64 1:0.1.40-11.el7_8 extras 43 k\n criu x86_64 3.12-2.el7 base 453 k\n fuse-overlayfs x86_64 0.7.2-6.el7_8 extras 54 k\n fuse3-libs x86_64 3.6.1-4.el7 extras 82 k\n libnet x86_64 1.1.6-7.el7 base 59 k\n libnftnl x86_64 1.0.8-3.el7 base 78 k\n nftables x86_64 1:0.8-14.el7 base 186 k\n protobuf-c x86_64 1.0.2-3.el7 base 28 k\n runc x86_64 1.0.0-70.rc10.el7_9 extras 2.7 M\n slirp4netns x86_64 0.4.3-4.el7_8 extras 81 k\n\nTransaction Summary\n================================================================================\nInstall 1 Package (+13 Dependent packages)\n\nTotal download size: 37 M\nInstalled size: 143 M\nDownloading packages:\n--------------------------------------------------------------------------------\nTotal 30 MB/s | 37 MB 00:01 \nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n Installing : 2:container-selinux-2.119.2-1.911c772.el7_8.noarch 1/14 \n Installing : slirp4netns-0.4.3-4.el7_8.x86_64 2/14 \n Installing : containernetworking-plugins-0.8.3-3.el7.centos.x86_64 3/14 \n Installing : libnftnl-1.0.8-3.el7.x86_64 4/14 \n Installing : 1:nftables-0.8-14.el7.x86_64 5/14 \n Installing : 2:conmon-2.0.8-1.el7.x86_64 6/14 \n Installing : fuse3-libs-3.6.1-4.el7.x86_64 7/14 \n Installing : fuse-overlayfs-0.7.2-6.el7_8.x86_64 8/14 \n Installing : 1:containers-common-0.1.40-11.el7_8.x86_64 9/14 \n Installing : protobuf-c-1.0.2-3.el7.x86_64 10/14 \n Installing : libnet-1.1.6-7.el7.x86_64 11/14 \n Installing : criu-3.12-2.el7.x86_64 12/14 \n Installing : runc-1.0.0-70.rc10.el7_9.x86_64 13/14 \n Installing : podman-1.6.4-36.el7_9.x86_64 14/14 \n Verifying : libnet-1.1.6-7.el7.x86_64 1/14 \n Verifying : protobuf-c-1.0.2-3.el7.x86_64 2/14 \n Verifying : fuse3-libs-3.6.1-4.el7.x86_64 3/14 \n Verifying : podman-1.6.4-36.el7_9.x86_64 4/14 \n Verifying : fuse-overlayfs-0.7.2-6.el7_8.x86_64 5/14 \n Verifying : runc-1.0.0-70.rc10.el7_9.x86_64 6/14 \n Verifying : slirp4netns-0.4.3-4.el7_8.x86_64 7/14 \n Verifying : 1:nftables-0.8-14.el7.x86_64 8/14 \n Verifying : criu-3.12-2.el7.x86_64 9/14 \n Verifying : 2:conmon-2.0.8-1.el7.x86_64 10/14 \n Verifying : 1:containers-common-0.1.40-11.el7_8.x86_64 11/14 \n Verifying : libnftnl-1.0.8-3.el7.x86_64 12/14 \n Verifying : containernetworking-plugins-0.8.3-3.el7.centos.x86_64 13/14 \n Verifying : 2:container-selinux-2.119.2-1.911c772.el7_8.noarch 14/14 \n\nInstalled:\n podman.x86_64 0:1.6.4-36.el7_9 \n\nDependency Installed:\n conmon.x86_64 2:2.0.8-1.el7 \n container-selinux.noarch 2:2.119.2-1.911c772.el7_8 \n containernetworking-plugins.x86_64 0:0.8.3-3.el7.centos \n containers-common.x86_64 1:0.1.40-11.el7_8 \n criu.x86_64 0:3.12-2.el7 \n fuse-overlayfs.x86_64 0:0.7.2-6.el7_8 \n fuse3-libs.x86_64 0:3.6.1-4.el7 \n libnet.x86_64 0:1.1.6-7.el7 \n libnftnl.x86_64 0:1.0.8-3.el7 \n nftables.x86_64 1:0.8-14.el7 \n protobuf-c.x86_64 0:1.0.2-3.el7 \n runc.x86_64 0:1.0.0-70.rc10.el7_9 \n slirp4netns.x86_64 0:0.4.3-4.el7_8 \n\nComplete!\n" ] } lsrpackages: podman TASK [Clean up Candlepin container] ******************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:38 Saturday 21 March 2026 15:33:17 -0400 (0:00:18.907) 0:00:21.318 ******** included: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/teardown_candlepin.yml for managed-node1 TASK [Check if the candlepin container exists] ********************************* task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/teardown_candlepin.yml:6 Saturday 21 March 2026 15:33:17 -0400 (0:00:00.048) 0:00:21.367 ******** ok: [managed-node1] => { "changed": false, "cmd": [ "podman", "ps", "-a", "--filter", "name=candlepin" ], "delta": "0:00:01.888599", "end": "2026-03-21 15:33:19.576353", "rc": 0, "start": "2026-03-21 15:33:17.687754" } STDOUT: CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES TASK [Ensure that Candlepin container doesn't exist] *************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/teardown_candlepin.yml:17 Saturday 21 March 2026 15:33:19 -0400 (0:00:02.329) 0:00:23.696 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Start Candlepin container] *********************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:41 Saturday 21 March 2026 15:33:19 -0400 (0:00:00.037) 0:00:23.734 ******** ok: [managed-node1] => { "changed": false, "cmd": [ "podman", "run", "--rm", "--detach", "--hostname", "candlepin.local", "--name", "candlepin", "--publish", "8443:8443", "--publish", "8080:8080", "--privileged", "ghcr.io/candlepin/candlepin-unofficial" ], "delta": "0:00:24.704517", "end": "2026-03-21 15:33:44.629309", "rc": 0, "start": "2026-03-21 15:33:19.924792" } STDOUT: d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa STDERR: Trying to pull ghcr.io/candlepin/candlepin-unofficial... Getting image source signatures Copying blob sha256:5baae3f93712d079b6030b8c02b29acecd6a7a6cdce52ab304b31425a048be6b Copying config sha256:6c8d0128d946443dc2cb0b755129351b01ff7b7c65670349e7d53b40a05309c5 Writing manifest to image destination Storing signatures TASK [Ensure directories exist] ************************************************ task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:62 Saturday 21 March 2026 15:33:44 -0400 (0:00:25.032) 0:00:48.766 ******** ok: [managed-node1] => (item=/etc/pki/product) => { "ansible_loop_var": "item", "changed": false, "gid": 0, "group": "root", "item": "/etc/pki/product", "mode": "0755", "owner": "root", "path": "/etc/pki/product", "secontext": "unconfined_u:object_r:cert_t:s0", "size": 4096, "state": "directory", "uid": 0 } ok: [managed-node1] => (item=/etc/pki/product-default) => { "ansible_loop_var": "item", "changed": false, "gid": 0, "group": "root", "item": "/etc/pki/product-default", "mode": "0755", "owner": "root", "path": "/etc/pki/product-default", "secontext": "unconfined_u:object_r:cert_t:s0", "size": 4096, "state": "directory", "uid": 0 } ok: [managed-node1] => (item=/etc/rhsm/ca) => { "ansible_loop_var": "item", "changed": false, "gid": 0, "group": "root", "item": "/etc/rhsm/ca", "mode": "0755", "owner": "root", "path": "/etc/rhsm/ca", "secontext": "system_u:object_r:rhsmcertd_config_t:s0", "size": 4096, "state": "directory", "uid": 0 } TASK [Copy product certificates] *********************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:72 Saturday 21 March 2026 15:33:46 -0400 (0:00:02.131) 0:00:50.897 ******** ok: [managed-node1] => (item=7050) => { "ansible_loop_var": "item", "changed": false, "cmd": [ "podman", "cp", "candlepin:/home/candlepin/devel/candlepin/generated_certs/7050.pem", "/etc/pki/product-default/" ], "delta": "0:00:00.523905", "end": "2026-03-21 15:33:48.015364", "item": "7050", "rc": 0, "start": "2026-03-21 15:33:47.491459" } TASK [Copy Candlepin CA certificate for subscription-manager] ****************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:83 Saturday 21 March 2026 15:33:48 -0400 (0:00:01.322) 0:00:52.220 ******** ok: [managed-node1] => { "changed": false, "cmd": [ "podman", "cp", "candlepin:/etc/candlepin/certs/candlepin-ca.crt", "/etc/rhsm/ca/candlepin-ca.pem" ], "delta": "0:00:00.517069", "end": "2026-03-21 15:33:49.459371", "rc": 0, "start": "2026-03-21 15:33:48.942302" } TASK [Copy Candlepin CA certificate for system] ******************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:92 Saturday 21 March 2026 15:33:49 -0400 (0:00:01.437) 0:00:53.658 ******** ok: [managed-node1] => { "changed": false, "cmd": [ "podman", "cp", "candlepin:/etc/candlepin/certs/candlepin-ca.crt", "/etc/pki/ca-trust/source/anchors/candlepin-ca.pem" ], "delta": "0:00:00.542804", "end": "2026-03-21 15:33:50.955601", "rc": 0, "start": "2026-03-21 15:33:50.412797" } TASK [Update system certificates store] **************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:101 Saturday 21 March 2026 15:33:51 -0400 (0:00:01.506) 0:00:55.164 ******** ok: [managed-node1] => { "changed": false, "cmd": [ "update-ca-trust", "extract" ], "delta": "0:00:02.029300", "end": "2026-03-21 15:33:53.982965", "rc": 0, "start": "2026-03-21 15:33:51.953665" } TASK [Wait for started Candlepin] ********************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:108 Saturday 21 March 2026 15:33:54 -0400 (0:00:03.057) 0:00:58.221 ******** ok: [managed-node1] => { "attempts": 1, "changed": false, "connection": "close", "content_type": "application/json", "cookies": {}, "cookies_string": "", "date": "Sat, 21 Mar 2026 19:34:13 GMT", "elapsed": 19, "redirected": true, "status": 200, "transfer_encoding": "chunked", "url": "https://candlepin.local:8443/candlepin/", "vary": "accept-encoding", "x_candlepin_request_uuid": "f4614d32-09a5-45d7-9331-698b3ac17065", "x_version": "4.7.3-1" } MSG: OK (unknown bytes) TASK [Install GPG key for RPM repositories] ************************************ task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:118 Saturday 21 March 2026 15:34:14 -0400 (0:00:20.144) 0:01:18.366 ******** changed: [managed-node1] => { "changed": true, "checksum_dest": null, "checksum_src": "e535dabdc941afb531fa9bb75b9a98d22bca8b81", "dest": "/etc/pki/rpm-gpg/RPM-GPG-KEY-candlepin", "elapsed": 0, "gid": 0, "group": "root", "md5sum": "eeaf1f5c1d5537f19a46506be9014ae6", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:cert_t:s0", "size": 1660, "src": "/root/.ansible/tmp/ansible-tmp-1774121654.36-12925-143798607795944/tmpKnyLxt", "state": "file", "status_code": 200, "uid": 0, "url": "http://candlepin.local:8080/RPM-GPG-KEY-candlepin" } MSG: OK (1660 bytes) TASK [Add environments] ******************************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:127 Saturday 21 March 2026 15:34:14 -0400 (0:00:00.546) 0:01:18.912 ******** skipping: [managed-node1] => (item={u'id': u'envId1', u'name': u'Environment 1', u'desc': u'The environment 1'}) => { "ansible_loop_var": "item", "changed": false, "item": { "desc": "The environment 1", "id": "envId1", "name": "Environment 1" }, "skip_reason": "Conditional result was False" } skipping: [managed-node1] => (item={u'id': u'envId2', u'name': u'Environment 2', u'desc': u'The environment 2'}) => { "ansible_loop_var": "item", "changed": false, "item": { "desc": "The environment 2", "id": "envId2", "name": "Environment 2" }, "skip_reason": "Conditional result was False" } TASK [Check Candlepin works] *************************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/check_candlepin.yml:3 Saturday 21 March 2026 15:34:14 -0400 (0:00:00.054) 0:01:18.967 ******** ok: [managed-node1] => { "changed": false, "connection": "close", "content_type": "application/json", "cookies": {}, "cookies_string": "", "date": "Sat, 21 Mar 2026 19:34:15 GMT", "elapsed": 0, "redirected": true, "status": 200, "transfer_encoding": "chunked", "url": "https://candlepin.local:8443/candlepin/", "vary": "accept-encoding", "x_candlepin_request_uuid": "e4a5e046-4a04-4f6d-9a76-f428fc5cfaa1", "x_version": "4.7.3-1" } MSG: OK (unknown bytes) TASK [Install packages for squid] ********************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml:7 Saturday 21 March 2026 15:34:15 -0400 (0:00:00.450) 0:01:19.417 ******** changed: [managed-node1] => { "changed": true, "changes": { "installed": [ "squid", "httpd-tools" ] }, "rc": 0, "results": [ "Loaded plugins: fastestmirror, product-id, search-disabled-repos, subscription-\n : manager\n\nThis system is not registered with an entitlement server. You can use subscription-manager to register.\n\nLoading mirror speeds from cached hostfile\nResolving Dependencies\n--> Running transaction check\n---> Package httpd-tools.x86_64 0:2.4.6-99.el7.centos.1 will be installed\n--> Processing Dependency: libaprutil-1.so.0()(64bit) for package: httpd-tools-2.4.6-99.el7.centos.1.x86_64\n--> Processing Dependency: libapr-1.so.0()(64bit) for package: httpd-tools-2.4.6-99.el7.centos.1.x86_64\n---> Package squid.x86_64 7:3.5.20-17.el7_9.10 will be installed\n--> Processing Dependency: squid-migration-script for package: 7:squid-3.5.20-17.el7_9.10.x86_64\n--> Processing Dependency: perl(DBI) for package: 7:squid-3.5.20-17.el7_9.10.x86_64\n--> Processing Dependency: libltdl.so.7()(64bit) for package: 7:squid-3.5.20-17.el7_9.10.x86_64\n--> Processing Dependency: libecap.so.3()(64bit) for package: 7:squid-3.5.20-17.el7_9.10.x86_64\n--> Running transaction check\n---> Package apr.x86_64 0:1.4.8-7.el7 will be installed\n---> Package apr-util.x86_64 0:1.5.2-6.el7_9.1 will be installed\n---> Package libecap.x86_64 0:1.0.0-1.el7 will be installed\n---> Package libtool-ltdl.x86_64 0:2.4.2-22.el7_3 will be installed\n---> Package perl-DBI.x86_64 0:1.627-4.el7 will be installed\n--> Processing Dependency: perl(RPC::PlServer) >= 0.2001 for package: perl-DBI-1.627-4.el7.x86_64\n--> Processing Dependency: perl(RPC::PlClient) >= 0.2000 for package: perl-DBI-1.627-4.el7.x86_64\n---> Package squid-migration-script.x86_64 7:3.5.20-17.el7_9.10 will be installed\n--> Running transaction check\n---> Package perl-PlRPC.noarch 0:0.2020-14.el7 will be installed\n--> Processing Dependency: perl(Net::Daemon) >= 0.13 for package: perl-PlRPC-0.2020-14.el7.noarch\n--> Processing Dependency: perl(Net::Daemon::Test) for package: perl-PlRPC-0.2020-14.el7.noarch\n--> Processing Dependency: perl(Net::Daemon::Log) for package: perl-PlRPC-0.2020-14.el7.noarch\n--> Running transaction check\n---> Package perl-Net-Daemon.noarch 0:0.48-5.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package Arch Version Repository Size\n================================================================================\nInstalling:\n httpd-tools x86_64 2.4.6-99.el7.centos.1 updates 94 k\n squid x86_64 7:3.5.20-17.el7_9.10 updates 3.1 M\nInstalling for dependencies:\n apr x86_64 1.4.8-7.el7 base 104 k\n apr-util x86_64 1.5.2-6.el7_9.1 updates 92 k\n libecap x86_64 1.0.0-1.el7 base 21 k\n libtool-ltdl x86_64 2.4.2-22.el7_3 base 49 k\n perl-DBI x86_64 1.627-4.el7 base 802 k\n perl-Net-Daemon noarch 0.48-5.el7 base 51 k\n perl-PlRPC noarch 0.2020-14.el7 base 36 k\n squid-migration-script x86_64 7:3.5.20-17.el7_9.10 updates 52 k\n\nTransaction Summary\n================================================================================\nInstall 2 Packages (+8 Dependent packages)\n\nTotal download size: 4.4 M\nInstalled size: 13 M\nDownloading packages:\n--------------------------------------------------------------------------------\nTotal 6.3 MB/s | 4.4 MB 00:00 \nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n Installing : apr-1.4.8-7.el7.x86_64 1/10 \n Installing : apr-util-1.5.2-6.el7_9.1.x86_64 2/10 \n Installing : libtool-ltdl-2.4.2-22.el7_3.x86_64 3/10 \n Installing : libecap-1.0.0-1.el7.x86_64 4/10 \n Installing : 7:squid-migration-script-3.5.20-17.el7_9.10.x86_64 5/10 \n Installing : perl-Net-Daemon-0.48-5.el7.noarch 6/10 \n Installing : perl-PlRPC-0.2020-14.el7.noarch 7/10 \n Installing : perl-DBI-1.627-4.el7.x86_64 8/10 \n Installing : 7:squid-3.5.20-17.el7_9.10.x86_64 9/10 \n Installing : httpd-tools-2.4.6-99.el7.centos.1.x86_64 10/10 \n Verifying : perl-Net-Daemon-0.48-5.el7.noarch 1/10 \n Verifying : httpd-tools-2.4.6-99.el7.centos.1.x86_64 2/10 \n Verifying : 7:squid-migration-script-3.5.20-17.el7_9.10.x86_64 3/10 \n Verifying : apr-util-1.5.2-6.el7_9.1.x86_64 4/10 \n Verifying : apr-1.4.8-7.el7.x86_64 5/10 \n Verifying : libecap-1.0.0-1.el7.x86_64 6/10 \n Verifying : libtool-ltdl-2.4.2-22.el7_3.x86_64 7/10 \n Verifying : 7:squid-3.5.20-17.el7_9.10.x86_64 8/10 \n Verifying : perl-DBI-1.627-4.el7.x86_64 9/10 \n Verifying : perl-PlRPC-0.2020-14.el7.noarch 10/10 \n\nInstalled:\n httpd-tools.x86_64 0:2.4.6-99.el7.centos.1 squid.x86_64 7:3.5.20-17.el7_9.10 \n\nDependency Installed:\n apr.x86_64 0:1.4.8-7.el7 \n apr-util.x86_64 0:1.5.2-6.el7_9.1 \n libecap.x86_64 0:1.0.0-1.el7 \n libtool-ltdl.x86_64 0:2.4.2-22.el7_3 \n perl-DBI.x86_64 0:1.627-4.el7 \n perl-Net-Daemon.noarch 0:0.48-5.el7 \n perl-PlRPC.noarch 0:0.2020-14.el7 \n squid-migration-script.x86_64 7:3.5.20-17.el7_9.10 \n\nComplete!\n" ] } lsrpackages: httpd-tools squid TASK [Check the status of the backup of configuration] ************************* task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml:15 Saturday 21 March 2026 15:34:21 -0400 (0:00:05.876) 0:01:25.294 ******** ok: [managed-node1] => { "changed": false, "stat": { "exists": false } } TASK [Backup the configuration] ************************************************ task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml:20 Saturday 21 March 2026 15:34:21 -0400 (0:00:00.295) 0:01:25.590 ******** changed: [managed-node1] => { "changed": true, "checksum": "70f953b530665efdfdd23b298cf8eddbacda74c0", "dest": "/etc/squid/squid.conf.BACKUP", "gid": 0, "group": "root", "md5sum": "3c8af1986cc652e972c35db57e808d36", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:squid_conf_t:s0", "size": 2244, "src": "/etc/squid/squid.conf", "state": "file", "uid": 0 } TASK [Copy the pristine configuration back] ************************************ task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml:29 Saturday 21 March 2026 15:34:21 -0400 (0:00:00.453) 0:01:26.043 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Open the Candlepin port] ************************************************* task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml:38 Saturday 21 March 2026 15:34:22 -0400 (0:00:00.038) 0:01:26.082 ******** changed: [managed-node1] => { "backup": "", "changed": true } MSG: line added TASK [Set the shutdown lifetime] *********************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml:47 Saturday 21 March 2026 15:34:22 -0400 (0:00:00.322) 0:01:26.404 ******** changed: [managed-node1] => { "backup": "", "changed": true } MSG: line added TASK [Set the port] ************************************************************ task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml:57 Saturday 21 March 2026 15:34:22 -0400 (0:00:00.344) 0:01:26.749 ******** ok: [managed-node1] => { "backup": "", "changed": false } TASK [Create the new passwd file] ********************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml:66 Saturday 21 March 2026 15:34:22 -0400 (0:00:00.303) 0:01:27.052 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set the port] ************************************************************ task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml:78 Saturday 21 March 2026 15:34:23 -0400 (0:00:00.038) 0:01:27.090 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Disable HTTP access allow] *********************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml:84 Saturday 21 March 2026 15:34:23 -0400 (0:00:00.038) 0:01:27.129 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Insert initial auth config] ********************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml:90 Saturday 21 March 2026 15:34:23 -0400 (0:00:00.037) 0:01:27.166 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Add authenticated acl] *************************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml:103 Saturday 21 March 2026 15:34:23 -0400 (0:00:00.037) 0:01:27.203 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Allow authenticated acl] ************************************************* task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml:111 Saturday 21 March 2026 15:34:23 -0400 (0:00:00.037) 0:01:27.241 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Restart squid] *********************************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml:119 Saturday 21 March 2026 15:34:23 -0400 (0:00:00.038) 0:01:27.279 ******** changed: [managed-node1] => { "changed": true, "name": "squid", "state": "started", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "system.slice network.target syslog.target nss-lookup.target systemd-journald.socket basic.target", "AllowIsolate": "no", "AmbientCapabilities": "0", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "18446744073709551615", "CPUAccounting": "no", "CPUQuotaPerSecUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "18446744073709551615", "CanIsolate": "no", "CanReload": "yes", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "18446744073709551615", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "Conflicts": "shutdown.target", "ControlPID": "0", "DefaultDependencies": "yes", "Delegate": "no", "Description": "Squid caching proxy", "DevicePolicy": "auto", "EnvironmentFile": "/etc/sysconfig/squid (ignore_errors=no)", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecReload": "{ path=/usr/sbin/squid ; argv[]=/usr/sbin/squid $SQUID_OPTS -k reconfigure -f $SQUID_CONF ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStart": "{ path=/usr/sbin/squid ; argv[]=/usr/sbin/squid $SQUID_OPTS -f $SQUID_CONF ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartPre": "{ path=/usr/libexec/squid/cache_swap.sh ; argv[]=/usr/libexec/squid/cache_swap.sh ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/sbin/squid ; argv[]=/usr/sbin/squid -k shutdown -f $SQUID_CONF ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/squid.service", "GuessMainPID": "yes", "IOScheduling": "0", "Id": "squid.service", "IgnoreOnIsolate": "no", "IgnoreOnSnapshot": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobTimeoutAction": "none", "JobTimeoutUSec": "0", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "18446744073709551615", "LimitCORE": "18446744073709551615", "LimitCPU": "18446744073709551615", "LimitDATA": "18446744073709551615", "LimitFSIZE": "18446744073709551615", "LimitLOCKS": "18446744073709551615", "LimitMEMLOCK": "65536", "LimitMSGQUEUE": "819200", "LimitNICE": "0", "LimitNOFILE": "16384", "LimitNPROC": "14311", "LimitRSS": "18446744073709551615", "LimitRTPRIO": "0", "LimitRTTIME": "18446744073709551615", "LimitSIGPENDING": "14311", "LimitSTACK": "18446744073709551615", "LoadState": "loaded", "MainPID": "0", "MemoryAccounting": "no", "MemoryCurrent": "18446744073709551615", "MemoryLimit": "18446744073709551615", "MountFlags": "0", "Names": "squid.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PermissionsStartOnly": "no", "PrivateDevices": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "ProtectHome": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "Requires": "system.slice basic.target", "Restart": "no", "RestartUSec": "100ms", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitInterval": "10000000", "StartupBlockIOWeight": "18446744073709551615", "StartupCPUShares": "18446744073709551615", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "no", "TasksCurrent": "18446744073709551615", "TasksMax": "18446744073709551615", "TimeoutStartUSec": "0", "TimeoutStopUSec": "0", "TimerSlackNSec": "50000", "Transient": "no", "Type": "forking", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "disabled", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "0" } } TASK [Add SELinux policy for proxy ports] ************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tests_proxy.yml:25 Saturday 21 March 2026 15:34:23 -0400 (0:00:00.714) 0:01:27.993 ******** ERROR! the role 'fedora.linux_system_roles.selinux' was not found in fedora.linux_system_roles:ansible.legacy:/tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/roles:/root/.ansible/roles:/usr/share/ansible/roles:/etc/ansible/roles:/tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc The error appears to be in '/tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tests_proxy.yml': line 27, column 19, but may be elsewhere in the file depending on the exact syntax problem. The offending line appears to be: include_role: name: fedora.linux_system_roles.selinux ^ here TASK [Unregister] ************************************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tests_proxy.yml:342 Saturday 21 March 2026 15:34:23 -0400 (0:00:00.050) 0:01:28.044 ******** included: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/run_role_with_clear_facts.yml for managed-node1 META: facts cleared TASK [Run the role] ************************************************************ task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/run_role_with_clear_facts.yml:22 Saturday 21 March 2026 15:34:24 -0400 (0:00:00.083) 0:01:28.127 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Run the role normally] *************************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/run_role_with_clear_facts.yml:32 Saturday 21 March 2026 15:34:24 -0400 (0:00:00.036) 0:01:28.163 ******** TASK [fedora.linux_system_roles.rhc : Set ansible_facts required by role] ****** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/main.yml:3 Saturday 21 March 2026 15:34:24 -0400 (0:00:00.064) 0:01:28.228 ******** included: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml for managed-node1 TASK [fedora.linux_system_roles.rhc : Ensure ansible_facts used by role] ******* task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml:3 Saturday 21 March 2026 15:34:24 -0400 (0:00:00.071) 0:01:28.300 ******** ok: [managed-node1] TASK [fedora.linux_system_roles.rhc : Check if system is ostree] *************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml:11 Saturday 21 March 2026 15:34:24 -0400 (0:00:00.406) 0:01:28.706 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Set flag to indicate system is ostree] *** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml:16 Saturday 21 March 2026 15:34:24 -0400 (0:00:00.017) 0:01:28.724 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Check if insights-packages are installed] *** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml:20 Saturday 21 March 2026 15:34:24 -0400 (0:00:00.015) 0:01:28.740 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Handle insights unregistration] ********** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/main.yml:6 Saturday 21 March 2026 15:34:24 -0400 (0:00:00.017) 0:01:28.757 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Handle system subscription] ************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/main.yml:15 Saturday 21 March 2026 15:34:24 -0400 (0:00:00.015) 0:01:28.773 ******** included: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml for managed-node1 TASK [fedora.linux_system_roles.rhc : Ensure required packages are installed] *** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:3 Saturday 21 March 2026 15:34:24 -0400 (0:00:00.031) 0:01:28.804 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Get subscription status] ***************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:10 Saturday 21 March 2026 15:34:24 -0400 (0:00:00.025) 0:01:28.830 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Call subscription-manager] *************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:23 Saturday 21 March 2026 15:34:24 -0400 (0:00:00.017) 0:01:28.847 ******** ok: [managed-node1] => { "changed": false } MSG: System already unregistered. TASK [fedora.linux_system_roles.rhc : Set or unset the release] **************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:49 Saturday 21 March 2026 15:34:25 -0400 (0:00:00.602) 0:01:29.450 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.rhc : Configure repositories] ****************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:58 Saturday 21 March 2026 15:34:25 -0400 (0:00:00.016) 0:01:29.467 ******** TASK [fedora.linux_system_roles.rhc : Handle insights registration] ************ task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/main.yml:18 Saturday 21 March 2026 15:34:25 -0400 (0:00:00.014) 0:01:29.481 ******** skipping: [managed-node1] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up Candlepin container] ******************************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tests_proxy.yml:347 Saturday 21 March 2026 15:34:25 -0400 (0:00:00.017) 0:01:29.498 ******** included: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/teardown_candlepin.yml for managed-node1 TASK [Check if the candlepin container exists] ********************************* task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/teardown_candlepin.yml:6 Saturday 21 March 2026 15:34:25 -0400 (0:00:00.058) 0:01:29.556 ******** ok: [managed-node1] => { "changed": false, "cmd": [ "podman", "ps", "-a", "--filter", "name=candlepin" ], "delta": "0:00:00.081441", "end": "2026-03-21 15:34:25.805407", "rc": 0, "start": "2026-03-21 15:34:25.723966" } STDOUT: CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES d46fc5f260ff ghcr.io/candlepin/candlepin-unofficial:latest /sbin/init 43 seconds ago Up 41 seconds ago 0.0.0.0:8080->8080/tcp candlepin TASK [Ensure that Candlepin container doesn't exist] *************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/teardown_candlepin.yml:17 Saturday 21 March 2026 15:34:25 -0400 (0:00:00.359) 0:01:29.916 ******** changed: [managed-node1] => { "changed": true, "cmd": [ "podman", "stop", "candlepin" ], "delta": "0:00:00.626987", "end": "2026-03-21 15:34:26.715005", "rc": 0, "start": "2026-03-21 15:34:26.088018" } STDOUT: d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa TASK [Remove SELinux policy for proxy ports] *********************************** task path: /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tests_proxy.yml:350 Saturday 21 March 2026 15:34:26 -0400 (0:00:00.918) 0:01:30.834 ******** ERROR! the role 'fedora.linux_system_roles.selinux' was not found in fedora.linux_system_roles:ansible.legacy:/tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/roles:/root/.ansible/roles:/usr/share/ansible/roles:/etc/ansible/roles:/tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc The error appears to be in '/tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tests_proxy.yml': line 352, column 19, but may be elsewhere in the file depending on the exact syntax problem. The offending line appears to be: include_role: name: fedora.linux_system_roles.selinux ^ here PLAY RECAP ********************************************************************* managed-node1 : ok=38 changed=9 unreachable=0 failed=0 skipped=23 rescued=0 ignored=0 SYSTEM ROLES ERRORS BEGIN v1 [] SYSTEM ROLES ERRORS END v1 TASKS RECAP ******************************************************************** Saturday 21 March 2026 15:34:26 -0400 (0:00:00.066) 0:01:30.901 ******** =============================================================================== Start Candlepin container ---------------------------------------------- 25.03s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:41 Wait for started Candlepin --------------------------------------------- 20.14s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:108 Install needed packages ------------------------------------------------ 18.91s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:31 Install packages for squid ---------------------------------------------- 5.88s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml:7 Update system certificates store ---------------------------------------- 3.06s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:101 Check if the candlepin container exists --------------------------------- 2.33s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/teardown_candlepin.yml:6 Ensure directories exist ------------------------------------------------ 2.13s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:62 Copy Candlepin CA certificate for system -------------------------------- 1.51s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:92 Copy Candlepin CA certificate for subscription-manager ------------------ 1.44s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:83 Copy product certificates ----------------------------------------------- 1.32s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:72 Ensure that Candlepin container doesn't exist --------------------------- 0.92s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/teardown_candlepin.yml:17 Gathering Facts --------------------------------------------------------- 0.82s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tests_proxy.yml:3 Restart squid ----------------------------------------------------------- 0.71s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml:119 fedora.linux_system_roles.rhc : Call subscription-manager --------------- 0.60s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/subscription-manager.yml:23 Install GPG key for RPM repositories ------------------------------------ 0.55s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:118 Backup the configuration ------------------------------------------------ 0.45s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_squid.yml:20 Check Candlepin works --------------------------------------------------- 0.45s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/check_candlepin.yml:3 Check if system is ostree ----------------------------------------------- 0.41s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_test_data.yml:32 fedora.linux_system_roles.rhc : Ensure ansible_facts used by role ------- 0.41s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/roles/rhc/tasks/set_vars.yml:3 Add candlepin hostname to /etc/hosts ------------------------------------ 0.38s /tmp/collections-6AW/ansible_collections/fedora/linux_system_roles/tests/rhc/tasks/setup_candlepin.yml:25 -- Logs begin at Sat 2026-03-21 15:27:15 EDT, end at Sat 2026-03-21 15:34:27 EDT. -- Mar 21 15:32:56 managed-node1 sudo[10542]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-ryjmgvqgvbvnkazfdxewopeqormzoacf ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121576.0-11631-8351668824400/AnsiballZ_setup.py Mar 21 15:32:56 managed-node1 sudo[10542]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:32:56 managed-node1 ansible-setup[10545]: Invoked with filter=* gather_subset=['all'] fact_path=/etc/ansible/facts.d gather_timeout=10 Mar 21 15:32:56 managed-node1 sudo[10542]: pam_unix(sudo:session): session closed for user root Mar 21 15:32:57 managed-node1 sudo[10619]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-vewmwhbelaztvyxkwrsfendbthuhsibm ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121577.0-11662-133551696335648/AnsiballZ_stat.py Mar 21 15:32:57 managed-node1 sudo[10619]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:32:57 managed-node1 ansible-stat[10622]: Invoked with checksum_algorithm=sha1 get_checksum=True follow=False path=/run/ostree-booted get_md5=False get_mime=True get_attributes=True Mar 21 15:32:57 managed-node1 sudo[10619]: pam_unix(sudo:session): session closed for user root Mar 21 15:32:57 managed-node1 sudo[10671]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-mnlqvddgagxeypiodvyehnuxtxweahli ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121577.59-11688-199501768892137/AnsiballZ_setup.py Mar 21 15:32:57 managed-node1 sudo[10671]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:32:57 managed-node1 ansible-setup[10674]: Invoked with filter=* gather_subset=['!all', '!min', 'distribution', 'distribution_major_version'] fact_path=/etc/ansible/facts.d gather_timeout=10 Mar 21 15:32:57 managed-node1 sudo[10671]: pam_unix(sudo:session): session closed for user root Mar 21 15:32:58 managed-node1 sudo[10727]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-hlxkvaqkueeawnazhbgkrmltizvujnnr ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121578.02-11701-163105044175623/AnsiballZ_lineinfile.py Mar 21 15:32:58 managed-node1 sudo[10727]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:32:58 managed-node1 ansible-lineinfile[10730]: Invoked with directory_mode=None force=None remote_src=None backrefs=False insertafter=None path=/etc/hosts owner=None follow=False validate=None group=None insertbefore=None unsafe_writes=False create=False state=present content=NOT_LOGGING_PARAMETER serole=None setype=None selevel=None regexp=.*candlepin.local line=127.0.0.1 candlepin.local src=None seuser=None delimiter=None mode=None firstmatch=False attributes=None backup=False Mar 21 15:32:58 managed-node1 sudo[10727]: pam_unix(sudo:session): session closed for user root Mar 21 15:32:58 managed-node1 sudo[10779]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-jcwkbubyuwogxupwkjdvvbwcvpyrbpml ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121578.46-11717-17141488419423/AnsiballZ_yum.py Mar 21 15:32:58 managed-node1 sudo[10779]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:32:58 managed-node1 ansible-yum[10782]: Invoked with lock_timeout=30 update_cache=False disable_excludes=None exclude=[] allow_downgrade=False disable_gpg_check=False conf_file=None use_backend=auto state=present disablerepo=[] releasever=None skip_broken=False autoremove=False download_dir=None enable_plugin=[] installroot=/ install_weak_deps=True name=['podman'] download_only=False bugfix=False list=None install_repoquery=True update_only=False disable_plugin=[] enablerepo=[] security=False validate_certs=True Mar 21 15:33:03 managed-node1 dbus[504]: [system] Reloaded configuration Mar 21 15:33:03 managed-node1 setsebool[10823]: The virt_use_nfs policy boolean was changed to 1 by root Mar 21 15:33:03 managed-node1 setsebool[10823]: The virt_sandbox_use_all_caps policy boolean was changed to 1 by root Mar 21 15:33:10 managed-node1 kernel: SELinux: 2048 avtab hash slots, 113365 rules. Mar 21 15:33:10 managed-node1 kernel: SELinux: 2048 avtab hash slots, 113365 rules. Mar 21 15:33:10 managed-node1 kernel: SELinux: 8 users, 14 roles, 5054 types, 318 bools, 1 sens, 1024 cats Mar 21 15:33:10 managed-node1 kernel: SELinux: 130 classes, 113365 rules Mar 21 15:33:10 managed-node1 kernel: SELinux: Converting 2283 SID table entries... Mar 21 15:33:12 managed-node1 dbus[504]: [system] Reloaded configuration Mar 21 15:33:12 managed-node1 yum[10793]: Installed: 2:container-selinux-2.119.2-1.911c772.el7_8.noarch Mar 21 15:33:12 managed-node1 yum[10793]: Installed: slirp4netns-0.4.3-4.el7_8.x86_64 Mar 21 15:33:14 managed-node1 yum[10793]: Installed: containernetworking-plugins-0.8.3-3.el7.centos.x86_64 Mar 21 15:33:14 managed-node1 yum[10793]: Installed: libnftnl-1.0.8-3.el7.x86_64 Mar 21 15:33:14 managed-node1 systemd[1]: Reloading. Mar 21 15:33:14 managed-node1 yum[10793]: Installed: 1:nftables-0.8-14.el7.x86_64 Mar 21 15:33:14 managed-node1 yum[10793]: Installed: 2:conmon-2.0.8-1.el7.x86_64 Mar 21 15:33:14 managed-node1 yum[10793]: Installed: fuse3-libs-3.6.1-4.el7.x86_64 Mar 21 15:33:14 managed-node1 yum[10793]: Installed: fuse-overlayfs-0.7.2-6.el7_8.x86_64 Mar 21 15:33:14 managed-node1 yum[10793]: Installed: 1:containers-common-0.1.40-11.el7_8.x86_64 Mar 21 15:33:14 managed-node1 yum[10793]: Installed: protobuf-c-1.0.2-3.el7.x86_64 Mar 21 15:33:14 managed-node1 yum[10793]: Installed: libnet-1.1.6-7.el7.x86_64 Mar 21 15:33:14 managed-node1 yum[10793]: Installed: criu-3.12-2.el7.x86_64 Mar 21 15:33:15 managed-node1 yum[10793]: Installed: runc-1.0.0-70.rc10.el7_9.x86_64 Mar 21 15:33:16 managed-node1 yum[10793]: Installed: podman-1.6.4-36.el7_9.x86_64 Mar 21 15:33:17 managed-node1 sudo[10779]: pam_unix(sudo:session): session closed for user root Mar 21 15:33:17 managed-node1 sudo[10913]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-jlccopxattyjoytzawehdolafsapyoyu ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121597.36-11854-25957538714312/AnsiballZ_command.py Mar 21 15:33:17 managed-node1 sudo[10913]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:33:17 managed-node1 ansible-command[10916]: Invoked with creates=None executable=None _uses_shell=False strip_empty_ends=True _raw_params=None removes=None argv=['podman', 'ps', '-a', '--filter', 'name=candlepin'] warn=True chdir=None stdin_add_newline=True stdin=None Mar 21 15:33:17 managed-node1 kernel: TECH PREVIEW: Overlay filesystem may not be fully supported. Please review provided documentation for limitations. Mar 21 15:33:19 managed-node1 podman[10917]: 2026-03-21 15:33:19.56176898 -0400 EDT m=+1.849225942 system refresh Mar 21 15:33:19 managed-node1 sudo[10913]: pam_unix(sudo:session): session closed for user root Mar 21 15:33:19 managed-node1 sudo[10984]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-tbyobnraalvcnhbeccdmypepbpttwwnr ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121599.73-11886-253407878375571/AnsiballZ_command.py Mar 21 15:33:19 managed-node1 sudo[10984]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:33:19 managed-node1 ansible-command[10987]: Invoked with creates=None executable=None _uses_shell=False strip_empty_ends=True _raw_params=None removes=None argv=['podman', 'run', '--rm', '--detach', '--hostname', 'candlepin.local', '--name', 'candlepin', '--publish', '8443:8443', '--publish', '8080:8080', '--privileged', 'ghcr.io/candlepin/candlepin-unofficial'] warn=True chdir=None stdin_add_newline=True stdin=None Mar 21 15:33:42 managed-node1 podman[10988]: 2026-03-21 15:33:42.191910189 -0400 EDT m=+22.245922113 image pull Mar 21 15:33:42 managed-node1 podman[10988]: 2026-03-21 15:33:42.215753251 -0400 EDT m=+22.269764830 container create d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa (image=ghcr.io/candlepin/candlepin-unofficial:latest, name=candlepin) Mar 21 15:33:42 managed-node1 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 21 15:33:42 managed-node1 kernel: IPv6: ADDRCONF(NETDEV_UP): eth0: link is not ready Mar 21 15:33:42 managed-node1 kernel: IPv6: ADDRCONF(NETDEV_UP): veth78740d52: link is not ready Mar 21 15:33:42 managed-node1 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): veth78740d52: link becomes ready Mar 21 15:33:42 managed-node1 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Mar 21 15:33:42 managed-node1 NetworkManager[572]: [1774121622.6846] manager: (cni-podman0): new Bridge device (/org/freedesktop/NetworkManager/Devices/3) Mar 21 15:33:42 managed-node1 NetworkManager[572]: [1774121622.6858] device (veth78740d52): carrier: link connected Mar 21 15:33:42 managed-node1 NetworkManager[572]: [1774121622.6861] manager: (veth78740d52): new Veth device (/org/freedesktop/NetworkManager/Devices/4) Mar 21 15:33:42 managed-node1 kernel: cni-podman0: port 1(veth78740d52) entered blocking state Mar 21 15:33:42 managed-node1 kernel: cni-podman0: port 1(veth78740d52) entered disabled state Mar 21 15:33:42 managed-node1 kernel: device veth78740d52 entered promiscuous mode Mar 21 15:33:42 managed-node1 kernel: cni-podman0: port 1(veth78740d52) entered blocking state Mar 21 15:33:42 managed-node1 kernel: cni-podman0: port 1(veth78740d52) entered forwarding state Mar 21 15:33:42 managed-node1 NetworkManager[572]: [1774121622.6975] device (cni-podman0): carrier: link connected Mar 21 15:33:42 managed-node1 NetworkManager[572]: [1774121622.6978] device (cni-podman0): state change: unmanaged -> unavailable (reason 'connection-assumed', sys-iface-state: 'external') Mar 21 15:33:42 managed-node1 NetworkManager[572]: [1774121622.7011] ifcfg-rh: add connection in-memory (99ca39c4-7668-451a-811b-21365618a92c,"cni-podman0") Mar 21 15:33:42 managed-node1 NetworkManager[572]: [1774121622.7018] device (cni-podman0): state change: unavailable -> disconnected (reason 'connection-assumed', sys-iface-state: 'external') Mar 21 15:33:42 managed-node1 NetworkManager[572]: [1774121622.7030] device (cni-podman0): Activation: starting connection 'cni-podman0' (99ca39c4-7668-451a-811b-21365618a92c) Mar 21 15:33:42 managed-node1 NetworkManager[572]: [1774121622.7032] device (cni-podman0): state change: disconnected -> prepare (reason 'none', sys-iface-state: 'external') Mar 21 15:33:42 managed-node1 NetworkManager[572]: [1774121622.7036] device (cni-podman0): state change: prepare -> config (reason 'none', sys-iface-state: 'external') Mar 21 15:33:42 managed-node1 NetworkManager[572]: [1774121622.7040] device (cni-podman0): state change: config -> ip-config (reason 'none', sys-iface-state: 'external') Mar 21 15:33:42 managed-node1 NetworkManager[572]: [1774121622.7046] device (cni-podman0): state change: ip-config -> ip-check (reason 'none', sys-iface-state: 'external') Mar 21 15:33:42 managed-node1 NetworkManager[572]: [1774121622.7050] device (cni-podman0): state change: ip-check -> secondaries (reason 'none', sys-iface-state: 'external') Mar 21 15:33:42 managed-node1 NetworkManager[572]: [1774121622.7052] device (cni-podman0): state change: secondaries -> activated (reason 'none', sys-iface-state: 'external') Mar 21 15:33:43 managed-node1 kernel: nf_conntrack version 0.5.0 (16384 buckets, 65536 max) Mar 21 15:33:43 managed-node1 NetworkManager[572]: [1774121623.3082] device (cni-podman0): Activation: successful, device activated. Mar 21 15:33:43 managed-node1 dbus[504]: [system] Activating via systemd: service name='org.freedesktop.nm_dispatcher' unit='dbus-org.freedesktop.nm-dispatcher.service' Mar 21 15:33:43 managed-node1 systemd[1]: Starting Network Manager Script Dispatcher Service... -- Subject: Unit NetworkManager-dispatcher.service has begun start-up -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- -- Unit NetworkManager-dispatcher.service has begun starting up. Mar 21 15:33:43 managed-node1 dbus[504]: [system] Successfully activated service 'org.freedesktop.nm_dispatcher' Mar 21 15:33:43 managed-node1 nm-dispatcher[11044]: req:1 'up' [cni-podman0]: new request (4 scripts) Mar 21 15:33:43 managed-node1 systemd[1]: Started Network Manager Script Dispatcher Service. -- Subject: Unit NetworkManager-dispatcher.service has finished start-up -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- -- Unit NetworkManager-dispatcher.service has finished starting up. -- -- The start-up result is done. Mar 21 15:33:43 managed-node1 nm-dispatcher[11044]: req:1 'up' [cni-podman0]: start running ordered scripts... Mar 21 15:33:44 managed-node1 systemd[1]: Created slice Virtual Machine and Container Slice. -- Subject: Unit machine.slice has finished start-up -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- -- Unit machine.slice has finished starting up. -- -- The start-up result is done. Mar 21 15:33:44 managed-node1 systemd[1]: Started libpod-conmon-d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa.scope. -- Subject: Unit libpod-conmon-d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa.scope has finished start-up -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- -- Unit libpod-conmon-d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa.scope has finished starting up. -- -- The start-up result is done. Mar 21 15:33:44 managed-node1 systemd[1]: Started libcontainer container d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa. -- Subject: Unit libpod-d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa.scope has finished start-up -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- -- Unit libpod-d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa.scope has finished starting up. -- -- The start-up result is done. Mar 21 15:33:44 managed-node1 kernel: SELinux: mount invalid. Same superblock, different security settings for (dev mqueue, type mqueue) Mar 21 15:33:44 managed-node1 podman[10988]: 2026-03-21 15:33:44.609747847 -0400 EDT m=+24.663759520 container init d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa (image=ghcr.io/candlepin/candlepin-unofficial:latest, name=candlepin) Mar 21 15:33:44 managed-node1 podman[10988]: 2026-03-21 15:33:44.624130156 -0400 EDT m=+24.678142116 container start d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa (image=ghcr.io/candlepin/candlepin-unofficial:latest, name=candlepin) Mar 21 15:33:44 managed-node1 sudo[10984]: pam_unix(sudo:session): session closed for user root Mar 21 15:33:44 managed-node1 systemd-rc-local-generator[10]: /etc/rc.d/rc.local is not marked executable, skipping. Mar 21 15:33:44 managed-node1 systemd-journald[19]: Received client request to flush runtime journal. Mar 21 15:33:45 managed-node1 systemd[1]: getty@tty1.service has no holdoff time, scheduling restart. Mar 21 15:33:45 managed-node1 systemd[1]: Stopped Getty on tty1. -- Subject: Unit getty@tty1.service has finished shutting down -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- -- Unit getty@tty1.service has finished shutting down. Mar 21 15:33:45 managed-node1 systemd[1]: Started Getty on tty1. -- Subject: Unit getty@tty1.service has finished start-up -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- -- Unit getty@tty1.service has finished starting up. -- -- The start-up result is done. Mar 21 15:33:45 managed-node1 sudo[11336]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-ntqcrjdobopkzbbwiifngmsnqqedyhpn ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121624.77-12131-237246772054976/AnsiballZ_file.py Mar 21 15:33:45 managed-node1 sudo[11336]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:33:45 managed-node1 ansible-file[11339]: Invoked with directory_mode=None force=False remote_src=None _original_basename=None path=/etc/pki/product owner=None follow=True group=None unsafe_writes=False state=directory content=NOT_LOGGING_PARAMETER serole=None selevel=None setype=None access_time=None access_time_format=%Y%m%d%H%M.%S modification_time=None regexp=None src=None seuser=None recurse=False _diff_peek=None delimiter=None mode=0755 modification_time_format=%Y%m%d%H%M.%S attributes=None backup=None Mar 21 15:33:45 managed-node1 sudo[11336]: pam_unix(sudo:session): session closed for user root Mar 21 15:33:45 managed-node1 sudo[11391]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-kjsbklhntexpkkvxsomsmtjrblndcoht ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121625.46-12131-260376495194051/AnsiballZ_file.py Mar 21 15:33:45 managed-node1 sudo[11391]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:33:45 managed-node1 ansible-file[11394]: Invoked with directory_mode=None force=False remote_src=None _original_basename=None path=/etc/pki/product-default owner=None follow=True group=None unsafe_writes=False state=directory content=NOT_LOGGING_PARAMETER serole=None selevel=None setype=None access_time=None access_time_format=%Y%m%d%H%M.%S modification_time=None regexp=None src=None seuser=None recurse=False _diff_peek=None delimiter=None mode=0755 modification_time_format=%Y%m%d%H%M.%S attributes=None backup=None Mar 21 15:33:46 managed-node1 sudo[11391]: pam_unix(sudo:session): session closed for user root Mar 21 15:33:46 managed-node1 sudo[11443]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-ybppmidhdzpguazwpdymxoehfiptcfjs ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121626.16-12131-154592295951924/AnsiballZ_file.py Mar 21 15:33:46 managed-node1 sudo[11443]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:33:46 managed-node1 ansible-file[11446]: Invoked with directory_mode=None force=False remote_src=None _original_basename=None path=/etc/rhsm/ca owner=None follow=True group=None unsafe_writes=False state=directory content=NOT_LOGGING_PARAMETER serole=None selevel=None setype=None access_time=None access_time_format=%Y%m%d%H%M.%S modification_time=None regexp=None src=None seuser=None recurse=False _diff_peek=None delimiter=None mode=0755 modification_time_format=%Y%m%d%H%M.%S attributes=None backup=None Mar 21 15:33:46 managed-node1 sudo[11443]: pam_unix(sudo:session): session closed for user root Mar 21 15:33:47 managed-node1 sudo[11495]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-mjbyawqtedqydiphuphyowppwkqcthsa ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121626.94-12168-249672360170003/AnsiballZ_command.py Mar 21 15:33:47 managed-node1 sudo[11495]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:33:47 managed-node1 ansible-command[11498]: Invoked with creates=None executable=None _uses_shell=False strip_empty_ends=True _raw_params=None removes=None argv=['podman', 'cp', 'candlepin:/home/candlepin/devel/candlepin/generated_certs/7050.pem', '/etc/pki/product-default/'] warn=True chdir=None stdin_add_newline=True stdin=None Mar 21 15:33:47 managed-node1 podman[11499]: 2026-03-21 15:33:47.643196253 -0400 EDT m=+0.101894986 container mount d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa (image=ghcr.io/candlepin/candlepin-unofficial:latest, name=candlepin) Mar 21 15:33:47 managed-node1 podman[11499]: 2026-03-21 15:33:47.667233587 -0400 EDT m=+0.125932065 container pause d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa (image=ghcr.io/candlepin/candlepin-unofficial:latest, name=candlepin) Mar 21 15:33:48 managed-node1 podman[11499]: 2026-03-21 15:33:48.00816076 -0400 EDT m=+0.466859882 container unpause d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa (image=ghcr.io/candlepin/candlepin-unofficial:latest, name=candlepin) Mar 21 15:33:48 managed-node1 podman[11499]: 2026-03-21 15:33:48.008689898 -0400 EDT m=+0.467388651 container unmount d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa (image=ghcr.io/candlepin/candlepin-unofficial:latest, name=candlepin) Mar 21 15:33:48 managed-node1 sudo[11495]: pam_unix(sudo:session): session closed for user root Mar 21 15:33:48 managed-node1 sudo[11597]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-xultuntswitfzpoyxdhdbinlbcwqrevn ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121628.28-12200-49487245652600/AnsiballZ_command.py Mar 21 15:33:48 managed-node1 sudo[11597]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:33:48 managed-node1 ansible-command[11600]: Invoked with creates=None executable=None _uses_shell=False strip_empty_ends=True _raw_params=None removes=None argv=['podman', 'cp', 'candlepin:/etc/candlepin/certs/candlepin-ca.crt', '/etc/rhsm/ca/candlepin-ca.pem'] warn=True chdir=None stdin_add_newline=True stdin=None Mar 21 15:33:49 managed-node1 podman[11601]: 2026-03-21 15:33:49.081533727 -0400 EDT m=+0.095183053 container mount d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa (image=ghcr.io/candlepin/candlepin-unofficial:latest, name=candlepin) Mar 21 15:33:49 managed-node1 podman[11601]: 2026-03-21 15:33:49.115784864 -0400 EDT m=+0.129434000 container pause d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa (image=ghcr.io/candlepin/candlepin-unofficial:latest, name=candlepin) Mar 21 15:33:49 managed-node1 podman[11601]: 2026-03-21 15:33:49.448770901 -0400 EDT m=+0.462420173 container unpause d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa (image=ghcr.io/candlepin/candlepin-unofficial:latest, name=candlepin) Mar 21 15:33:49 managed-node1 podman[11601]: 2026-03-21 15:33:49.449311744 -0400 EDT m=+0.462961189 container unmount d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa (image=ghcr.io/candlepin/candlepin-unofficial:latest, name=candlepin) Mar 21 15:33:49 managed-node1 sudo[11597]: pam_unix(sudo:session): session closed for user root Mar 21 15:33:50 managed-node1 sudo[11700]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-bculxxuujybcpgzicdbxgpfeljjifdev ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121629.74-12225-248525150715664/AnsiballZ_command.py Mar 21 15:33:50 managed-node1 sudo[11700]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:33:50 managed-node1 ansible-command[11703]: Invoked with creates=None executable=None _uses_shell=False strip_empty_ends=True _raw_params=None removes=None argv=['podman', 'cp', 'candlepin:/etc/candlepin/certs/candlepin-ca.crt', '/etc/pki/ca-trust/source/anchors/candlepin-ca.pem'] warn=True chdir=None stdin_add_newline=True stdin=None Mar 21 15:33:50 managed-node1 podman[11704]: 2026-03-21 15:33:50.548934589 -0400 EDT m=+0.095604510 container mount d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa (image=ghcr.io/candlepin/candlepin-unofficial:latest, name=candlepin) Mar 21 15:33:50 managed-node1 podman[11704]: 2026-03-21 15:33:50.587771396 -0400 EDT m=+0.134441060 container pause d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa (image=ghcr.io/candlepin/candlepin-unofficial:latest, name=candlepin) Mar 21 15:33:50 managed-node1 podman[11704]: 2026-03-21 15:33:50.949162518 -0400 EDT m=+0.495832353 container unpause d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa (image=ghcr.io/candlepin/candlepin-unofficial:latest, name=candlepin) Mar 21 15:33:50 managed-node1 podman[11704]: 2026-03-21 15:33:50.949699556 -0400 EDT m=+0.496369349 container unmount d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa (image=ghcr.io/candlepin/candlepin-unofficial:latest, name=candlepin) Mar 21 15:33:50 managed-node1 sudo[11700]: pam_unix(sudo:session): session closed for user root Mar 21 15:33:51 managed-node1 sudo[11804]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-przavayelirkyiuxkdbngxlaqtgacnty ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121631.26-12276-255439742079817/AnsiballZ_command.py Mar 21 15:33:51 managed-node1 sudo[11804]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:33:51 managed-node1 ansible-command[11807]: Invoked with creates=None executable=None _uses_shell=False strip_empty_ends=True _raw_params=None removes=None argv=['update-ca-trust', 'extract'] warn=True chdir=None stdin_add_newline=True stdin=None Mar 21 15:33:53 managed-node1 sudo[11804]: pam_unix(sudo:session): session closed for user root Mar 21 15:33:54 managed-node1 sudo[11862]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-vqtjaoyerfrxwurzxwjdrauxawwlkpye ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121634.28-12368-250923817769560/AnsiballZ_uri.py Mar 21 15:33:54 managed-node1 sudo[11862]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:33:55 managed-node1 ansible-uri[11865]: Invoked with directory_mode=None force=False remote_src=None status_code=[200] body_format=raw owner=None follow=False client_key=None group=None use_proxy=True headers={} unsafe_writes=False serole=None content=NOT_LOGGING_PARAMETER setype=None follow_redirects=safe return_content=False client_cert=None body=None timeout=30 src=None dest=None selevel=None force_basic_auth=False removes=None http_agent=ansible-httpget regexp=None url_password=NOT_LOGGING_PARAMETER url=https://candlepin.local:8443/candlepin validate_certs=False seuser=None method=HEAD creates=None unix_socket=None delimiter=None mode=None url_username=None attributes=None backup=None Mar 21 15:34:14 managed-node1 sudo[11862]: pam_unix(sudo:session): session closed for user root Mar 21 15:34:14 managed-node1 sudo[12017]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-xdrcvknlgxkdlmitfuavhyghdpozmmwp ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121654.36-12925-143798607795944/AnsiballZ_get_url.py Mar 21 15:34:14 managed-node1 sudo[12017]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:34:14 managed-node1 ansible-get_url[12020]: Invoked with directory_mode=None force=False remote_src=None owner=None follow=False client_key=None group=None use_proxy=True unsafe_writes=False serole=None content=NOT_LOGGING_PARAMETER validate_certs=True setype=None client_cert=None timeout=10 url_password=NOT_LOGGING_PARAMETER dest=/etc/pki/rpm-gpg/RPM-GPG-KEY-candlepin selevel=None force_basic_auth=False sha256sum= http_agent=ansible-httpget regexp=None src=None url=http://candlepin.local:8080/RPM-GPG-KEY-candlepin checksum= seuser=None headers=None delimiter=None mode=0644 url_username=None attributes=None backup=None tmp_dest=None Mar 21 15:34:14 managed-node1 sudo[12017]: pam_unix(sudo:session): session closed for user root Mar 21 15:34:15 managed-node1 sudo[12074]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-dmrivhugwxoaxtpsvrzdalrvuydmagiu ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121654.96-12951-91261377846918/AnsiballZ_uri.py Mar 21 15:34:15 managed-node1 sudo[12074]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:34:15 managed-node1 ansible-uri[12077]: Invoked with directory_mode=None force=False remote_src=None status_code=[200] body_format=raw owner=None follow=False client_key=None group=None use_proxy=True headers={} unsafe_writes=False serole=None content=NOT_LOGGING_PARAMETER setype=None follow_redirects=safe return_content=False client_cert=None body=None timeout=30 src=None dest=None selevel=None force_basic_auth=False removes=None http_agent=ansible-httpget regexp=None url_password=NOT_LOGGING_PARAMETER url=https://candlepin.local:8443/candlepin validate_certs=False seuser=None method=HEAD creates=None unix_socket=None delimiter=None mode=None url_username=None attributes=None backup=None Mar 21 15:34:15 managed-node1 sudo[12074]: pam_unix(sudo:session): session closed for user root Mar 21 15:34:15 managed-node1 sudo[12130]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-yqoiqfcuvjuiexawspelosxdtemsmjfp ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121655.47-12960-264297151083163/AnsiballZ_yum.py Mar 21 15:34:15 managed-node1 sudo[12130]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:34:15 managed-node1 ansible-yum[12133]: Invoked with lock_timeout=30 update_cache=False disable_excludes=None exclude=[] allow_downgrade=False disable_gpg_check=False conf_file=None use_backend=auto state=present disablerepo=[] releasever=None skip_broken=False autoremove=False download_dir=None enable_plugin=[] installroot=/ install_weak_deps=True name=['squid', 'httpd-tools'] download_only=False bugfix=False list=None install_repoquery=True update_only=False disable_plugin=[] enablerepo=[] security=False validate_certs=True Mar 21 15:34:19 managed-node1 yum[12144]: Installed: apr-1.4.8-7.el7.x86_64 Mar 21 15:34:19 managed-node1 yum[12144]: Installed: apr-util-1.5.2-6.el7_9.1.x86_64 Mar 21 15:34:19 managed-node1 yum[12144]: Installed: libtool-ltdl-2.4.2-22.el7_3.x86_64 Mar 21 15:34:19 managed-node1 yum[12144]: Installed: libecap-1.0.0-1.el7.x86_64 Mar 21 15:34:19 managed-node1 yum[12144]: Installed: 7:squid-migration-script-3.5.20-17.el7_9.10.x86_64 Mar 21 15:34:19 managed-node1 yum[12144]: Installed: perl-Net-Daemon-0.48-5.el7.noarch Mar 21 15:34:19 managed-node1 yum[12144]: Installed: perl-PlRPC-0.2020-14.el7.noarch Mar 21 15:34:19 managed-node1 yum[12144]: Installed: perl-DBI-1.627-4.el7.x86_64 Mar 21 15:34:19 managed-node1 groupadd[12176]: group added to /etc/group: name=squid, GID=23 Mar 21 15:34:19 managed-node1 groupadd[12176]: group added to /etc/gshadow: name=squid Mar 21 15:34:19 managed-node1 groupadd[12176]: new group: name=squid, GID=23 Mar 21 15:34:19 managed-node1 useradd[12181]: new user: name=squid, UID=23, GID=23, home=/var/spool/squid, shell=/sbin/nologin Mar 21 15:34:20 managed-node1 systemd[1]: Reloading. Mar 21 15:34:20 managed-node1 yum[12144]: Installed: 7:squid-3.5.20-17.el7_9.10.x86_64 Mar 21 15:34:20 managed-node1 yum[12144]: Installed: httpd-tools-2.4.6-99.el7.centos.1.x86_64 Mar 21 15:34:21 managed-node1 sudo[12130]: pam_unix(sudo:session): session closed for user root Mar 21 15:34:21 managed-node1 sudo[12252]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-shkwgtvdxttdkfqczpojasrztflztglp ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121661.29-13038-58546966089290/AnsiballZ_stat.py Mar 21 15:34:21 managed-node1 sudo[12252]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:34:21 managed-node1 ansible-stat[12255]: Invoked with checksum_algorithm=sha1 get_checksum=True follow=False path=/etc/squid/squid.conf.BACKUP get_md5=False get_mime=True get_attributes=True Mar 21 15:34:21 managed-node1 sudo[12252]: pam_unix(sudo:session): session closed for user root Mar 21 15:34:21 managed-node1 sudo[12304]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-uhxfndnzvgzowmbcchfdzxciceadytue ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121661.59-13047-29973629827020/AnsiballZ_copy.py Mar 21 15:34:21 managed-node1 sudo[12304]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:34:21 managed-node1 ansible-copy[12307]: Invoked with directory_mode=None force=True remote_src=True _original_basename=None owner=None follow=False local_follow=None group=None unsafe_writes=False setype=None content=NOT_LOGGING_PARAMETER serole=None dest=/etc/squid/squid.conf.BACKUP selevel=None regexp=None validate=None src=/etc/squid/squid.conf checksum=None seuser=None delimiter=None mode=0644 attributes=None backup=False Mar 21 15:34:21 managed-node1 sudo[12304]: pam_unix(sudo:session): session closed for user root Mar 21 15:34:22 managed-node1 sudo[12356]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-ltkbmcsadgvifwveywhbuxbuxierflue ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121662.09-13066-191892956151071/AnsiballZ_lineinfile.py Mar 21 15:34:22 managed-node1 sudo[12356]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:34:22 managed-node1 ansible-lineinfile[12359]: Invoked with directory_mode=None force=None remote_src=None backrefs=False owner=None path=/etc/squid/squid.conf insertafter=None follow=False validate=None group=None insertbefore=^acl Safe_ports unsafe_writes=False create=False state=present content=NOT_LOGGING_PARAMETER serole=None setype=None selevel=None regexp=^acl SSL_ports port 8443 line=acl SSL_ports port 8443 # Candlepin src=None seuser=None delimiter=None mode=None firstmatch=True attributes=None backup=False Mar 21 15:34:22 managed-node1 sudo[12356]: pam_unix(sudo:session): session closed for user root Mar 21 15:34:22 managed-node1 sudo[12408]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-admtrnetspbcyledkgfuusbeshrykxox ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121662.42-13079-268201920338464/AnsiballZ_lineinfile.py Mar 21 15:34:22 managed-node1 sudo[12408]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:34:22 managed-node1 ansible-lineinfile[12411]: Invoked with directory_mode=None force=None remote_src=None backrefs=False insertafter=None path=/etc/squid/squid.conf owner=None follow=False validate=None group=None insertbefore=None unsafe_writes=False create=False state=present content=NOT_LOGGING_PARAMETER serole=None setype=None selevel=None regexp=^shutdown_lifetime line=shutdown_lifetime 5 seconds src=None seuser=None delimiter=None mode=None firstmatch=False attributes=None backup=False Mar 21 15:34:22 managed-node1 sudo[12408]: pam_unix(sudo:session): session closed for user root Mar 21 15:34:22 managed-node1 sudo[12460]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-xbbbyedymadipmiswqsydhsudfumzply ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121662.75-13091-274576312433745/AnsiballZ_lineinfile.py Mar 21 15:34:22 managed-node1 sudo[12460]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:34:22 managed-node1 ansible-lineinfile[12463]: Invoked with directory_mode=None force=None remote_src=None backrefs=False insertafter=None path=/etc/squid/squid.conf owner=None follow=False validate=None group=None insertbefore=None unsafe_writes=False create=False state=present content=NOT_LOGGING_PARAMETER serole=None setype=None selevel=None regexp=^http_port line=http_port 3128 src=None seuser=None delimiter=None mode=None firstmatch=False attributes=None backup=False Mar 21 15:34:22 managed-node1 sudo[12460]: pam_unix(sudo:session): session closed for user root Mar 21 15:34:23 managed-node1 sudo[12512]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-knhybqzzshcwwdyugwqqpfmrcdnxrors ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121663.27-13120-244885246277571/AnsiballZ_systemd.py Mar 21 15:34:23 managed-node1 sudo[12512]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:34:23 managed-node1 ansible-systemd[12515]: Invoked with no_block=False force=None name=squid daemon_reexec=False enabled=None daemon_reload=False state=restarted masked=None scope=None user=None Mar 21 15:34:23 managed-node1 polkitd[499]: Registered Authentication Agent for unix-process:12519:42921 (system bus name :1.87 [/usr/bin/pkttyagent --notify-fd 5 --fallback], object path /org/freedesktop/PolicyKit1/AuthenticationAgent, locale en_US.UTF-8) Mar 21 15:34:23 managed-node1 systemd[1]: Starting Squid caching proxy... -- Subject: Unit squid.service has begun start-up -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- -- Unit squid.service has begun starting up. Mar 21 15:34:23 managed-node1 squid[12533]: Squid Parent: will start 1 kids Mar 21 15:34:23 managed-node1 squid[12533]: Squid Parent: (squid-1) process 12535 started Mar 21 15:34:23 managed-node1 systemd[1]: Started Squid caching proxy. -- Subject: Unit squid.service has finished start-up -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- -- Unit squid.service has finished starting up. -- -- The start-up result is done. Mar 21 15:34:23 managed-node1 sudo[12512]: pam_unix(sudo:session): session closed for user root Mar 21 15:34:23 managed-node1 polkitd[499]: Unregistered Authentication Agent for unix-process:12519:42921 (system bus name :1.87, object path /org/freedesktop/PolicyKit1/AuthenticationAgent, locale en_US.UTF-8) (disconnected from bus) Mar 21 15:34:24 managed-node1 sudo[12599]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-xbbdscedsoymabdalwweusyammkxwmgj ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121664.28-13146-231020368633065/AnsiballZ_setup.py Mar 21 15:34:24 managed-node1 sudo[12599]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:34:24 managed-node1 ansible-setup[12602]: Invoked with filter=* gather_subset=['!all', '!min', 'distribution', 'distribution_major_version', 'distribution_version'] fact_path=/etc/ansible/facts.d gather_timeout=10 Mar 21 15:34:24 managed-node1 sudo[12599]: pam_unix(sudo:session): session closed for user root Mar 21 15:34:25 managed-node1 sudo[12655]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-fjocvkthrmvvjjmsuwxqfbjlkfoymopc ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121664.85-13178-236109683215886/AnsiballZ_redhat_subscription.py Mar 21 15:34:25 managed-node1 sudo[12655]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:34:25 managed-node1 ansible-community.general.redhat_subscription[12658]: Invoked with server_hostname=None rhsm_repo_ca_cert=None consumer_id=None activationkey=NOT_LOGGING_PARAMETER server_proxy_password=NOT_LOGGING_PARAMETER consumer_name=None server_proxy_port=None server_proxy_scheme=None environment=None force_register=False state=absent server_proxy_user=None server_port=None username=None pool_ids=[] auto_attach=None server_insecure=None password=NOT_LOGGING_PARAMETER rhsm_baseurl=None consumer_type=None org_id=None syspurpose=None token=NOT_LOGGING_PARAMETER server_prefix=None release=None server_proxy_hostname=None Mar 21 15:34:25 managed-node1 sudo[12655]: pam_unix(sudo:session): session closed for user root Mar 21 15:34:25 managed-node1 sudo[12718]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-rytdauwednlswtlojxeoqeurvbhokdap ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121665.53-13195-25551300438133/AnsiballZ_command.py Mar 21 15:34:25 managed-node1 sudo[12718]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:34:25 managed-node1 ansible-command[12721]: Invoked with creates=None executable=None _uses_shell=False strip_empty_ends=True _raw_params=None removes=None argv=['podman', 'ps', '-a', '--filter', 'name=candlepin'] warn=True chdir=None stdin_add_newline=True stdin=None Mar 21 15:34:25 managed-node1 sudo[12718]: pam_unix(sudo:session): session closed for user root Mar 21 15:34:26 managed-node1 sudo[12778]: root : TTY=pts/0 ; PWD=/root ; USER=root ; COMMAND=/bin/sh -c echo BECOME-SUCCESS-qohldsepgmpmebsltdxakdacnixalvdw ; /usr/bin/python /root/.ansible/tmp/ansible-tmp-1774121665.89-13211-170427010850065/AnsiballZ_command.py Mar 21 15:34:26 managed-node1 sudo[12778]: pam_unix(sudo:session): session opened for user root by root(uid=0) Mar 21 15:34:26 managed-node1 ansible-command[12781]: Invoked with creates=None executable=None _uses_shell=False strip_empty_ends=True _raw_params=None removes=None argv=['podman', 'stop', 'candlepin'] warn=True chdir=None stdin_add_newline=True stdin=None Mar 21 15:34:26 managed-node1 systemd[1]: getty@tty1.service has no holdoff time, scheduling restart. Mar 21 15:34:26 managed-node1 systemd[1]: Stopped Getty on tty1. -- Subject: Unit getty@tty1.service has finished shutting down -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- -- Unit getty@tty1.service has finished shutting down. Mar 21 15:34:26 managed-node1 systemd[1]: Started Getty on tty1. -- Subject: Unit getty@tty1.service has finished start-up -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- -- Unit getty@tty1.service has finished starting up. -- -- The start-up result is done. Mar 21 15:34:26 managed-node1 systemd-journald[19]: Received SIGTERM from PID 1 (systemd-shutdow). Mar 21 15:34:26 managed-node1 podman[12782]: 2026-03-21 15:34:26.707688978 -0400 EDT m=+0.598603005 container died d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa (image=ghcr.io/candlepin/candlepin-unofficial:latest, name=candlepin) Mar 21 15:34:26 managed-node1 podman[12782]: 2026-03-21 15:34:26.711731301 -0400 EDT m=+0.602645336 container stop d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa (image=ghcr.io/candlepin/candlepin-unofficial:latest, name=candlepin) Mar 21 15:34:26 managed-node1 sudo[12778]: pam_unix(sudo:session): session closed for user root Mar 21 15:34:26 managed-node1 kernel: ip6_tables: (C) 2000-2006 Netfilter Core Team Mar 21 15:34:26 managed-node1 kernel: cni-podman0: port 1(veth78740d52) entered disabled state Mar 21 15:34:26 managed-node1 kernel: device veth78740d52 left promiscuous mode Mar 21 15:34:26 managed-node1 kernel: cni-podman0: port 1(veth78740d52) entered disabled state Mar 21 15:34:26 managed-node1 NetworkManager[572]: [1774121666.8320] device (veth78740d52): released from master device cni-podman0 Mar 21 15:34:26 managed-node1 podman[12840]: 2026-03-21 15:34:26.946361946 -0400 EDT m=+0.259392883 container remove d46fc5f260ffb2cac98ab89273e91b5bb53c32a59801b6d967f5c71216bbd9fa (image=ghcr.io/candlepin/candlepin-unofficial:latest, name=candlepin) Mar 21 15:34:27 managed-node1 sshd[12917]: Accepted publickey for root from 10.31.41.131 port 45002 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Mar 21 15:34:27 managed-node1 systemd[1]: Started Session 14 of user root. -- Subject: Unit session-14.scope has finished start-up -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- -- Unit session-14.scope has finished starting up. -- -- The start-up result is done. Mar 21 15:34:27 managed-node1 systemd-logind[525]: New session 14 of user root. -- Subject: A new session 14 has been created for user root -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- Documentation: http://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 14 has been created for the user root. -- -- The leading process of the session is 12917. Mar 21 15:34:27 managed-node1 sshd[12917]: pam_unix(sshd:session): session opened for user root by (uid=0) Mar 21 15:34:27 managed-node1 sshd[12917]: Received disconnect from 10.31.41.131 port 45002:11: disconnected by user Mar 21 15:34:27 managed-node1 sshd[12917]: Disconnected from 10.31.41.131 port 45002 Mar 21 15:34:27 managed-node1 sshd[12917]: pam_unix(sshd:session): session closed for user root Mar 21 15:34:27 managed-node1 systemd-logind[525]: Removed session 14. -- Subject: Session 14 has been terminated -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- Documentation: http://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A session with the ID 14 has been terminated. Mar 21 15:34:27 managed-node1 sshd[12934]: Accepted publickey for root from 10.31.41.131 port 45004 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Mar 21 15:34:27 managed-node1 systemd-logind[525]: New session 15 of user root. -- Subject: A new session 15 has been created for user root -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- Documentation: http://www.freedesktop.org/wiki/Software/systemd/multiseat -- -- A new session with the ID 15 has been created for the user root. -- -- The leading process of the session is 12934. Mar 21 15:34:27 managed-node1 systemd[1]: Started Session 15 of user root. -- Subject: Unit session-15.scope has finished start-up -- Defined-By: systemd -- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel -- -- Unit session-15.scope has finished starting up. -- -- The start-up result is done. Mar 21 15:34:27 managed-node1 sshd[12934]: pam_unix(sshd:session): session opened for user root by (uid=0)