ansible-playbook 2.9.27
  config file = /etc/ansible/ansible.cfg
  configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
  ansible python module location = /usr/lib/python2.7/site-packages/ansible
  executable location = /usr/bin/ansible-playbook
  python version = 2.7.5 (default, Nov 14 2023, 16:14:06) [GCC 4.8.5 20150623 (Red Hat 4.8.5-44)]
Using /etc/ansible/ansible.cfg as config file
[WARNING]: running playbook inside collection fedora.linux_system_roles
statically imported: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml
statically imported: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml
statically imported: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml
statically imported: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml
statically imported: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml
statically imported: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml
statically imported: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml
statically imported: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml
statically imported: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml
statically imported: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml
statically imported: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml
statically imported: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml
Skipping callback 'actionable', as we already have a stdout callback.
Skipping callback 'counter_enabled', as we already have a stdout callback.
Skipping callback 'debug', as we already have a stdout callback.
Skipping callback 'dense', as we already have a stdout callback.
Skipping callback 'dense', as we already have a stdout callback.
Skipping callback 'full_skip', as we already have a stdout callback.
Skipping callback 'json', as we already have a stdout callback.
Skipping callback 'minimal', as we already have a stdout callback.
Skipping callback 'null', as we already have a stdout callback.
Skipping callback 'oneline', as we already have a stdout callback.
Skipping callback 'selective', as we already have a stdout callback.
Skipping callback 'skippy', as we already have a stdout callback.
Skipping callback 'stderr', as we already have a stdout callback.
Skipping callback 'unixy', as we already have a stdout callback.
Skipping callback 'yaml', as we already have a stdout callback.

PLAYBOOK: tests_luks.yml *******************************************************
1 plays in /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml

PLAY [Test LUKS] ***************************************************************

TASK [Gathering Facts] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:2
Saturday 14 December 2024  17:54:12 -0500 (0:00:00.050)       0:00:00.050 ***** 
ok: [managed-node2]
META: ran handlers

TASK [Enable FIPS mode] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:20
Saturday 14 December 2024  17:54:13 -0500 (0:00:01.114)       0:00:01.164 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reboot] ******************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:28
Saturday 14 December 2024  17:54:13 -0500 (0:00:00.100)       0:00:01.265 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Enable FIPS mode] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:39
Saturday 14 December 2024  17:54:13 -0500 (0:00:00.075)       0:00:01.340 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reboot] ******************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:43
Saturday 14 December 2024  17:54:14 -0500 (0:00:00.073)       0:00:01.413 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Ensure dracut-fips] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:53
Saturday 14 December 2024  17:54:14 -0500 (0:00:00.071)       0:00:01.485 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Configure boot for FIPS] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:59
Saturday 14 December 2024  17:54:14 -0500 (0:00:00.070)       0:00:01.556 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reboot] ******************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:68
Saturday 14 December 2024  17:54:14 -0500 (0:00:00.070)       0:00:01.627 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Run the role] ************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:72
Saturday 14 December 2024  17:54:14 -0500 (0:00:00.063)       0:00:01.691 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  17:54:14 -0500 (0:00:00.099)       0:00:01.791 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  17:54:14 -0500 (0:00:00.052)       0:00:01.844 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  17:54:14 -0500 (0:00:00.056)       0:00:01.901 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  17:54:14 -0500 (0:00:00.089)       0:00:01.991 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "exists": false
    }
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  17:54:15 -0500 (0:00:00.549)       0:00:02.540 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "__storage_is_ostree": false
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  17:54:15 -0500 (0:00:00.078)       0:00:02.619 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  17:54:15 -0500 (0:00:00.033)       0:00:02.652 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  17:54:15 -0500 (0:00:00.035)       0:00:02.688 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  17:54:15 -0500 (0:00:00.106)       0:00:02.794 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  17:54:16 -0500 (0:00:01.463)       0:00:04.258 ***** 
ok: [managed-node2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  17:54:16 -0500 (0:00:00.082)       0:00:04.341 ***** 
ok: [managed-node2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  17:54:17 -0500 (0:00:00.087)       0:00:04.428 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  17:54:17 -0500 (0:00:00.826)       0:00:05.255 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  17:54:18 -0500 (0:00:00.140)       0:00:05.396 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  17:54:18 -0500 (0:00:00.053)       0:00:05.449 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  17:54:18 -0500 (0:00:00.080)       0:00:05.530 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  17:54:18 -0500 (0:00:00.037)       0:00:05.567 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: kpartx

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  17:54:18 -0500 (0:00:00.777)       0:00:06.344 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  17:54:20 -0500 (0:00:01.423)       0:00:07.767 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  17:54:20 -0500 (0:00:00.064)       0:00:07.832 ***** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  17:54:20 -0500 (0:00:00.037)       0:00:07.870 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Saturday 14 December 2024  17:54:20 -0500 (0:00:00.484)       0:00:08.354 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Saturday 14 December 2024  17:54:21 -0500 (0:00:00.032)       0:00:08.386 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734216835.4788961, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "72884e3f126482c2d28276ff7c57744fa95eff91", 
        "ctime": 1734216835.0358963, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 263588, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0644", 
        "mtime": 1734216835.0358963, 
        "nlink": 1, 
        "path": "/etc/fstab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 1229, 
        "uid": 0, 
        "version": "606402765", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Saturday 14 December 2024  17:54:21 -0500 (0:00:00.357)       0:00:08.744 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  17:54:21 -0500 (0:00:00.047)       0:00:08.791 ***** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Saturday 14 December 2024  17:54:21 -0500 (0:00:00.037)       0:00:08.829 ***** 
ok: [managed-node2] => {
    "blivet_output": {
        "actions": [], 
        "changed": false, 
        "crypts": [], 
        "failed": false, 
        "leaves": [], 
        "mounts": [], 
        "packages": [], 
        "pools": [], 
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Saturday 14 December 2024  17:54:21 -0500 (0:00:00.038)       0:00:08.867 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Saturday 14 December 2024  17:54:21 -0500 (0:00:00.038)       0:00:08.905 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Saturday 14 December 2024  17:54:21 -0500 (0:00:00.048)       0:00:08.953 ***** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Saturday 14 December 2024  17:54:21 -0500 (0:00:00.046)       0:00:09.000 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Saturday 14 December 2024  17:54:21 -0500 (0:00:00.052)       0:00:09.052 ***** 

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Saturday 14 December 2024  17:54:21 -0500 (0:00:00.085)       0:00:09.137 ***** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Saturday 14 December 2024  17:54:21 -0500 (0:00:00.040)       0:00:09.178 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Saturday 14 December 2024  17:54:21 -0500 (0:00:00.043)       0:00:09.222 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734216694.2399628, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", 
        "ctime": 1734216692.1449637, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 917509, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/x-empty", 
        "mode": "0600", 
        "mtime": 1734216692.1439638, 
        "nlink": 1, 
        "path": "/etc/crypttab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": false, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": "606407754", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Saturday 14 December 2024  17:54:22 -0500 (0:00:00.321)       0:00:09.543 ***** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Saturday 14 December 2024  17:54:22 -0500 (0:00:00.031)       0:00:09.574 ***** 
ok: [managed-node2]

TASK [Get unused disks] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:76
Saturday 14 December 2024  17:54:22 -0500 (0:00:00.719)       0:00:10.294 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml for managed-node2

TASK [Ensure test packages] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:2
Saturday 14 December 2024  17:54:22 -0500 (0:00:00.079)       0:00:10.373 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "util-linux-2.23.2-65.el7_9.1.x86_64 providing util-linux is already installed"
    ]
}
lsrpackages: util-linux

TASK [Find unused disks in the system] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:11
Saturday 14 December 2024  17:54:23 -0500 (0:00:00.672)       0:00:11.045 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "disks": [
        "sda"
    ], 
    "info": [
        "Line: NAME=\"/dev/sda\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"", 
        "Line: NAME=\"/dev/sdb\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"", 
        "Line: NAME=\"/dev/sdc\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"", 
        "Line: NAME=\"/dev/sdd\" TYPE=\"disk\" SIZE=\"1099511627776\" FSTYPE=\"\" LOG-SEC=\"512\"", 
        "Line: NAME=\"/dev/sde\" TYPE=\"disk\" SIZE=\"1099511627776\" FSTYPE=\"\" LOG-SEC=\"512\"", 
        "Line: NAME=\"/dev/sdf\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"", 
        "Line: NAME=\"/dev/sdg\" TYPE=\"disk\" SIZE=\"1099511627776\" FSTYPE=\"\" LOG-SEC=\"512\"", 
        "Line: NAME=\"/dev/sdh\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"", 
        "Line: NAME=\"/dev/sdi\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"", 
        "Line: NAME=\"/dev/xvda\" TYPE=\"disk\" SIZE=\"268435456000\" FSTYPE=\"\" LOG-SEC=\"512\"", 
        "Line: NAME=\"/dev/xvda1\" TYPE=\"part\" SIZE=\"268434390528\" FSTYPE=\"ext4\" LOG-SEC=\"512\"", 
        "Line type [part] is not disk: NAME=\"/dev/xvda1\" TYPE=\"part\" SIZE=\"268434390528\" FSTYPE=\"ext4\" LOG-SEC=\"512\"", 
        "filename [xvda1] is a partition", 
        "Disk [/dev/xvda] attrs [{'fstype': '', 'type': 'disk', 'ssize': '512', 'size': '268435456000'}] has partitions"
    ]
}

TASK [Debug why there are no unused disks] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:20
Saturday 14 December 2024  17:54:24 -0500 (0:00:00.656)       0:00:11.701 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set unused_disks if necessary] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:29
Saturday 14 December 2024  17:54:24 -0500 (0:00:00.063)       0:00:11.765 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "unused_disks": [
            "sda"
        ]
    }, 
    "changed": false
}

TASK [Exit playbook when there's not enough unused disks in the system] ********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:34
Saturday 14 December 2024  17:54:24 -0500 (0:00:00.114)       0:00:11.879 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Print unused disks] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:39
Saturday 14 December 2024  17:54:24 -0500 (0:00:00.050)       0:00:11.930 ***** 
ok: [managed-node2] => {
    "unused_disks": [
        "sda"
    ]
}

TASK [Test for correct handling of new encrypted volume w/ no key] *************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:85
Saturday 14 December 2024  17:54:24 -0500 (0:00:00.060)       0:00:11.990 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node2

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Saturday 14 December 2024  17:54:24 -0500 (0:00:00.123)       0:00:12.114 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_pools_global": [], 
        "storage_safe_mode_global": true, 
        "storage_volumes_global": []
    }, 
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Saturday 14 December 2024  17:54:24 -0500 (0:00:00.071)       0:00:12.186 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  17:54:24 -0500 (0:00:00.101)       0:00:12.287 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  17:54:25 -0500 (0:00:00.104)       0:00:12.392 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  17:54:25 -0500 (0:00:00.068)       0:00:12.460 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  17:54:25 -0500 (0:00:00.188)       0:00:12.656 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  17:54:25 -0500 (0:00:00.062)       0:00:12.719 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  17:54:25 -0500 (0:00:00.059)       0:00:12.778 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  17:54:25 -0500 (0:00:00.122)       0:00:12.901 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  17:54:25 -0500 (0:00:00.064)       0:00:12.966 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  17:54:25 -0500 (0:00:00.128)       0:00:13.094 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  17:54:26 -0500 (0:00:01.274)       0:00:14.369 ***** 
ok: [managed-node2] => {
    "storage_pools": []
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  17:54:27 -0500 (0:00:00.082)       0:00:14.451 ***** 
ok: [managed-node2] => {
    "storage_volumes": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": true, 
            "mount_point": "/opt/test1", 
            "name": "foo", 
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  17:54:27 -0500 (0:00:00.085)       0:00:14.537 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [
        "cryptsetup"
    ], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  17:54:31 -0500 (0:00:04.030)       0:00:18.568 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  17:54:31 -0500 (0:00:00.115)       0:00:18.683 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  17:54:31 -0500 (0:00:00.081)       0:00:18.764 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  17:54:31 -0500 (0:00:00.053)       0:00:18.817 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  17:54:31 -0500 (0:00:00.061)       0:00:18.879 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed", 
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: cryptsetup kpartx

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  17:54:32 -0500 (0:00:00.733)       0:00:19.612 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  17:54:33 -0500 (0:00:01.033)       0:00:20.646 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  17:54:33 -0500 (0:00:00.079)       0:00:20.725 ***** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  17:54:33 -0500 (0:00:00.049)       0:00:20.774 ***** 
fatal: [managed-node2]: FAILED! => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [], 
    "pools": [], 
    "volumes": []
}

MSG:

encrypted volume 'foo' missing key/password

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Saturday 14 December 2024  17:54:37 -0500 (0:00:03.935)       0:00:24.709 ***** 
fatal: [managed-node2]: FAILED! => {
    "changed": false
}

MSG:

{u'_ansible_no_log': False, u'crypts': [], u'pools': [], u'leaves': [], u'changed': False, u'actions': [], u'failed': True, u'volumes': [], u'invocation': {u'module_args': {u'packages_only': False, u'disklabel_type': None, u'diskvolume_mkfs_option_map': {u'ext4': u'-F', u'ext3': u'-F', u'ext2': u'-F'}, u'safe_mode': True, u'pools': [], u'volumes': [{u'raid_metadata_version': None, u'raid_level': None, u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': 10737418240, u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': None, u'encryption': True, u'mount_device_identifier': u'uuid', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'fs_overwrite_existing': True, u'encryption_key_size': None, u'deduplication': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'cache_size': 0, u'mount_user': None, u'raid_spare_count': None, u'cache_mode': None, u'name': u'foo', u'mount_group': None, u'type': u'disk', u'disks': [u'sda'], u'cached': False, u'thin_pool_size': None, u'thin': None, u'mount_check': 0, u'mount_passno': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}], u'pool_defaults': {u'raid_metadata_version': None, u'encryption_cipher': None, u'encryption_key': None, u'encryption_luks_version': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_password': None, u'encryption': False, u'disks': [], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [], u'shared': False, u'type': u'lvm', u'encryption_key_size': None, u'raid_chunk_size': None}, u'volume_defaults': {u'raid_metadata_version': None, u'raid_level': None, u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': 0, u'mount_point': u'', u'compression': None, u'encryption_password': None, u'encryption': False, u'mount_device_identifier': u'uuid', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'thin_pool_name': None, u'type': u'lvm', u'encryption_key_size': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'cache_size': 0, u'raid_spare_count': None, u'cache_mode': None, u'deduplication': None, u'cached': False, u'fs_overwrite_existing': True, u'disks': [], u'thin_pool_size': None, u'thin': None, u'mount_check': 0, u'mount_passno': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}, u'use_partitions': None}}, u'mounts': [], u'packages': [], u'msg': u"encrypted volume 'foo' missing key/password"}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  17:54:37 -0500 (0:00:00.058)       0:00:24.768 ***** 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Saturday 14 December 2024  17:54:37 -0500 (0:00:00.054)       0:00:24.822 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Saturday 14 December 2024  17:54:37 -0500 (0:00:00.049)       0:00:24.872 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Saturday 14 December 2024  17:54:37 -0500 (0:00:00.048)       0:00:24.920 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Create an encrypted disk volume w/ default fs] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:100
Saturday 14 December 2024  17:54:37 -0500 (0:00:00.033)       0:00:24.954 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  17:54:37 -0500 (0:00:00.115)       0:00:25.070 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  17:54:37 -0500 (0:00:00.075)       0:00:25.145 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  17:54:37 -0500 (0:00:00.051)       0:00:25.197 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  17:54:37 -0500 (0:00:00.100)       0:00:25.297 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  17:54:37 -0500 (0:00:00.035)       0:00:25.333 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  17:54:37 -0500 (0:00:00.032)       0:00:25.366 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  17:54:38 -0500 (0:00:00.033)       0:00:25.399 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  17:54:38 -0500 (0:00:00.033)       0:00:25.433 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  17:54:38 -0500 (0:00:00.112)       0:00:25.545 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  17:54:39 -0500 (0:00:01.543)       0:00:27.089 ***** 
ok: [managed-node2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  17:54:39 -0500 (0:00:00.037)       0:00:27.126 ***** 
ok: [managed-node2] => {
    "storage_volumes": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": true, 
            "encryption_password": "yabbadabbadoo", 
            "mount_point": "/opt/test1", 
            "name": "foo", 
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  17:54:39 -0500 (0:00:00.043)       0:00:27.170 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [
        "cryptsetup"
    ], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  17:54:43 -0500 (0:00:03.938)       0:00:31.109 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  17:54:43 -0500 (0:00:00.178)       0:00:31.287 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  17:54:43 -0500 (0:00:00.066)       0:00:31.354 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  17:54:44 -0500 (0:00:00.053)       0:00:31.408 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  17:54:44 -0500 (0:00:00.109)       0:00:31.517 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed", 
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: cryptsetup kpartx

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  17:54:45 -0500 (0:00:00.936)       0:00:32.454 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  17:54:46 -0500 (0:00:01.177)       0:00:33.632 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  17:54:46 -0500 (0:00:00.152)       0:00:33.785 ***** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  17:54:46 -0500 (0:00:00.063)       0:00:33.848 ***** 
changed: [managed-node2] => {
    "actions": [
        {
            "action": "create format", 
            "device": "/dev/sda", 
            "fs_type": "luks"
        }, 
        {
            "action": "create device", 
            "device": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
            "fs_type": null
        }, 
        {
            "action": "create format", 
            "device": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
            "fs_type": "xfs"
        }
    ], 
    "changed": true, 
    "crypts": [
        {
            "backing_device": "/dev/sda", 
            "name": "luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
            "password": "-", 
            "state": "present"
        }
    ], 
    "leaves": [
        "/dev/sdb", 
        "/dev/sdc", 
        "/dev/sdd", 
        "/dev/sde", 
        "/dev/sdf", 
        "/dev/sdg", 
        "/dev/sdh", 
        "/dev/sdi", 
        "/dev/xvda1", 
        "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269"
    ], 
    "mounts": [
        {
            "dump": 0, 
            "fstype": "xfs", 
            "group": null, 
            "mode": null, 
            "opts": "defaults", 
            "owner": null, 
            "passno": 0, 
            "path": "/opt/test1", 
            "src": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
            "state": "mounted"
        }
    ], 
    "packages": [
        "cryptsetup", 
        "xfsprogs", 
        "e2fsprogs"
    ], 
    "pools": [], 
    "volumes": [
        {
            "_device": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
            "_kernel_device": "/dev/dm-0", 
            "_mount_id": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
            "_raw_device": "/dev/sda", 
            "_raw_kernel_device": "/dev/sda", 
            "cache_devices": [], 
            "cache_mode": null, 
            "cache_size": 0, 
            "cached": false, 
            "compression": null, 
            "deduplication": null, 
            "disks": [
                "sda"
            ], 
            "encryption": true, 
            "encryption_cipher": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
            "fs_create_options": "", 
            "fs_label": "", 
            "fs_overwrite_existing": true, 
            "fs_type": "xfs", 
            "mount_check": 0, 
            "mount_device_identifier": "uuid", 
            "mount_group": null, 
            "mount_mode": null, 
            "mount_options": "defaults", 
            "mount_passno": 0, 
            "mount_point": "/opt/test1", 
            "mount_user": null, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "raid_stripe_size": null, 
            "size": 10737418240, 
            "state": "present", 
            "thin": null, 
            "thin_pool_name": null, 
            "thin_pool_size": null, 
            "type": "disk", 
            "vdo_pool_size": null
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Saturday 14 December 2024  17:54:57 -0500 (0:00:10.556)       0:00:44.405 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Saturday 14 December 2024  17:54:57 -0500 (0:00:00.051)       0:00:44.457 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734216835.4788961, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "72884e3f126482c2d28276ff7c57744fa95eff91", 
        "ctime": 1734216835.0358963, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 263588, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0644", 
        "mtime": 1734216835.0358963, 
        "nlink": 1, 
        "path": "/etc/fstab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 1229, 
        "uid": 0, 
        "version": "606402765", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Saturday 14 December 2024  17:54:57 -0500 (0:00:00.399)       0:00:44.857 ***** 
ok: [managed-node2] => {
    "backup": "", 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  17:54:58 -0500 (0:00:00.721)       0:00:45.579 ***** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Saturday 14 December 2024  17:54:58 -0500 (0:00:00.073)       0:00:45.652 ***** 
ok: [managed-node2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "create format", 
                "device": "/dev/sda", 
                "fs_type": "luks"
            }, 
            {
                "action": "create device", 
                "device": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
                "fs_type": null
            }, 
            {
                "action": "create format", 
                "device": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
                "fs_type": "xfs"
            }
        ], 
        "changed": true, 
        "crypts": [
            {
                "backing_device": "/dev/sda", 
                "name": "luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
                "password": "-", 
                "state": "present"
            }
        ], 
        "failed": false, 
        "leaves": [
            "/dev/sdb", 
            "/dev/sdc", 
            "/dev/sdd", 
            "/dev/sde", 
            "/dev/sdf", 
            "/dev/sdg", 
            "/dev/sdh", 
            "/dev/sdi", 
            "/dev/xvda1", 
            "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269"
        ], 
        "mounts": [
            {
                "dump": 0, 
                "fstype": "xfs", 
                "group": null, 
                "mode": null, 
                "opts": "defaults", 
                "owner": null, 
                "passno": 0, 
                "path": "/opt/test1", 
                "src": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
                "state": "mounted"
            }
        ], 
        "packages": [
            "cryptsetup", 
            "xfsprogs", 
            "e2fsprogs"
        ], 
        "pools": [], 
        "volumes": [
            {
                "_device": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
                "_kernel_device": "/dev/dm-0", 
                "_mount_id": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
                "_raw_device": "/dev/sda", 
                "_raw_kernel_device": "/dev/sda", 
                "cache_devices": [], 
                "cache_mode": null, 
                "cache_size": 0, 
                "cached": false, 
                "compression": null, 
                "deduplication": null, 
                "disks": [
                    "sda"
                ], 
                "encryption": true, 
                "encryption_cipher": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                "fs_create_options": "", 
                "fs_label": "", 
                "fs_overwrite_existing": true, 
                "fs_type": "xfs", 
                "mount_check": 0, 
                "mount_device_identifier": "uuid", 
                "mount_group": null, 
                "mount_mode": null, 
                "mount_options": "defaults", 
                "mount_passno": 0, 
                "mount_point": "/opt/test1", 
                "mount_user": null, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "raid_stripe_size": null, 
                "size": 10737418240, 
                "state": "present", 
                "thin": null, 
                "thin_pool_name": null, 
                "thin_pool_size": null, 
                "type": "disk", 
                "vdo_pool_size": null
            }
        ]
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Saturday 14 December 2024  17:54:58 -0500 (0:00:00.083)       0:00:45.736 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Saturday 14 December 2024  17:54:58 -0500 (0:00:00.057)       0:00:45.793 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": [
            {
                "_device": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
                "_kernel_device": "/dev/dm-0", 
                "_mount_id": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
                "_raw_device": "/dev/sda", 
                "_raw_kernel_device": "/dev/sda", 
                "cache_devices": [], 
                "cache_mode": null, 
                "cache_size": 0, 
                "cached": false, 
                "compression": null, 
                "deduplication": null, 
                "disks": [
                    "sda"
                ], 
                "encryption": true, 
                "encryption_cipher": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                "fs_create_options": "", 
                "fs_label": "", 
                "fs_overwrite_existing": true, 
                "fs_type": "xfs", 
                "mount_check": 0, 
                "mount_device_identifier": "uuid", 
                "mount_group": null, 
                "mount_mode": null, 
                "mount_options": "defaults", 
                "mount_passno": 0, 
                "mount_point": "/opt/test1", 
                "mount_user": null, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "raid_stripe_size": null, 
                "size": 10737418240, 
                "state": "present", 
                "thin": null, 
                "thin_pool_name": null, 
                "thin_pool_size": null, 
                "type": "disk", 
                "vdo_pool_size": null
            }
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Saturday 14 December 2024  17:54:58 -0500 (0:00:00.071)       0:00:45.864 ***** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Saturday 14 December 2024  17:54:58 -0500 (0:00:00.056)       0:00:45.921 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Saturday 14 December 2024  17:54:59 -0500 (0:00:01.215)       0:00:47.137 ***** 
changed: [managed-node2] => (item={u'src': u'/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": true, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
        "state": "mounted"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Saturday 14 December 2024  17:55:00 -0500 (0:00:00.742)       0:00:47.880 ***** 
skipping: [managed-node2] => (item={u'src': u'/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'})  => {
    "ansible_loop_var": "mount_info", 
    "changed": false, 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
        "state": "mounted"
    }, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Saturday 14 December 2024  17:55:00 -0500 (0:00:00.070)       0:00:47.951 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Saturday 14 December 2024  17:55:01 -0500 (0:00:00.568)       0:00:48.519 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734216694.2399628, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", 
        "ctime": 1734216692.1449637, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 917509, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/x-empty", 
        "mode": "0600", 
        "mtime": 1734216692.1439638, 
        "nlink": 1, 
        "path": "/etc/crypttab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": false, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": "606407754", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Saturday 14 December 2024  17:55:01 -0500 (0:00:00.482)       0:00:49.001 ***** 
changed: [managed-node2] => (item={u'state': u'present', u'password': u'-', u'name': u'luks-b0c50623-e6eb-411b-b5d3-25165f6ee269', u'backing_device': u'/dev/sda'}) => {
    "ansible_loop_var": "entry", 
    "backup": "", 
    "changed": true, 
    "entry": {
        "backing_device": "/dev/sda", 
        "name": "luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
        "password": "-", 
        "state": "present"
    }
}

MSG:

line added

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Saturday 14 December 2024  17:55:02 -0500 (0:00:00.488)       0:00:49.490 ***** 
ok: [managed-node2]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:112
Saturday 14 December 2024  17:55:03 -0500 (0:00:01.076)       0:00:50.567 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node2

TASK [Print out pool information] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Saturday 14 December 2024  17:55:03 -0500 (0:00:00.178)       0:00:50.746 ***** 
skipping: [managed-node2] => {}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Saturday 14 December 2024  17:55:03 -0500 (0:00:00.065)       0:00:50.811 ***** 
ok: [managed-node2] => {
    "_storage_volumes_list": [
        {
            "_device": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
            "_kernel_device": "/dev/dm-0", 
            "_mount_id": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
            "_raw_device": "/dev/sda", 
            "_raw_kernel_device": "/dev/sda", 
            "cache_devices": [], 
            "cache_mode": null, 
            "cache_size": 0, 
            "cached": false, 
            "compression": null, 
            "deduplication": null, 
            "disks": [
                "sda"
            ], 
            "encryption": true, 
            "encryption_cipher": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
            "fs_create_options": "", 
            "fs_label": "", 
            "fs_overwrite_existing": true, 
            "fs_type": "xfs", 
            "mount_check": 0, 
            "mount_device_identifier": "uuid", 
            "mount_group": null, 
            "mount_mode": null, 
            "mount_options": "defaults", 
            "mount_passno": 0, 
            "mount_point": "/opt/test1", 
            "mount_user": null, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "raid_stripe_size": null, 
            "size": 10737418240, 
            "state": "present", 
            "thin": null, 
            "thin_pool_name": null, 
            "thin_pool_size": null, 
            "type": "disk", 
            "vdo_pool_size": null
        }
    ]
}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Saturday 14 December 2024  17:55:03 -0500 (0:00:00.129)       0:00:50.941 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "info": {
        "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269": {
            "fstype": "xfs", 
            "label": "", 
            "mountpoint": "/opt/test1", 
            "name": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
            "size": "10G", 
            "type": "crypt", 
            "uuid": "d55b60e7-e805-4cd8-be19-51b765643608"
        }, 
        "/dev/sda": {
            "fstype": "crypto_LUKS", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sda", 
            "size": "10G", 
            "type": "disk", 
            "uuid": "b0c50623-e6eb-411b-b5d3-25165f6ee269"
        }, 
        "/dev/sdb": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdb", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdc": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdc", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdd": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdd", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sde": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sde", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdf": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdf", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdg": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdg", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdh": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdh", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdi": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdi", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/xvda", 
            "size": "250G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda1": {
            "fstype": "ext4", 
            "label": "", 
            "mountpoint": "/", 
            "name": "/dev/xvda1", 
            "size": "250G", 
            "type": "partition", 
            "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Saturday 14 December 2024  17:55:04 -0500 (0:00:00.597)       0:00:51.538 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/fstab"
    ], 
    "delta": "0:00:00.003122", 
    "end": "2024-12-14 17:55:04.670001", 
    "rc": 0, 
    "start": "2024-12-14 17:55:04.666879"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Thu Jun 20 10:23:46 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 /                       ext4    defaults        1 1
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269 /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Saturday 14 December 2024  17:55:04 -0500 (0:00:00.662)       0:00:52.200 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/crypttab"
    ], 
    "delta": "0:00:00.002783", 
    "end": "2024-12-14 17:55:05.268155", 
    "failed_when_result": false, 
    "rc": 0, 
    "start": "2024-12-14 17:55:05.265372"
}

STDOUT:

luks-b0c50623-e6eb-411b-b5d3-25165f6ee269 /dev/sda -

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Saturday 14 December 2024  17:55:05 -0500 (0:00:00.534)       0:00:52.735 ***** 

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Saturday 14 December 2024  17:55:05 -0500 (0:00:00.054)       0:00:52.789 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node2

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Saturday 14 December 2024  17:55:05 -0500 (0:00:00.114)       0:00:52.903 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true, 
        "_storage_volume_tests": [
            "mount", 
            "fstab", 
            "fs", 
            "device", 
            "encryption", 
            "md", 
            "size", 
            "cache"
        ]
    }, 
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Saturday 14 December 2024  17:55:05 -0500 (0:00:00.047)       0:00:52.951 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node2

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Saturday 14 December 2024  17:55:05 -0500 (0:00:00.238)       0:00:53.190 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269"
    }, 
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Saturday 14 December 2024  17:55:05 -0500 (0:00:00.064)       0:00:53.254 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1", 
        "storage_test_swap_expected_matches": "0"
    }, 
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Saturday 14 December 2024  17:55:05 -0500 (0:00:00.068)       0:00:53.323 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Saturday 14 December 2024  17:55:05 -0500 (0:00:00.053)       0:00:53.376 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Saturday 14 December 2024  17:55:06 -0500 (0:00:00.061)       0:00:53.437 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Saturday 14 December 2024  17:55:06 -0500 (0:00:00.054)       0:00:53.491 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Saturday 14 December 2024  17:55:06 -0500 (0:00:00.062)       0:00:53.554 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Saturday 14 December 2024  17:55:06 -0500 (0:00:00.054)       0:00:53.609 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Saturday 14 December 2024  17:55:06 -0500 (0:00:00.054)       0:00:53.663 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Saturday 14 December 2024  17:55:06 -0500 (0:00:00.052)       0:00:53.716 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Saturday 14 December 2024  17:55:06 -0500 (0:00:00.054)       0:00:53.771 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null, 
        "storage_test_mount_expected_mount_point": null, 
        "storage_test_swap_expected_matches": null, 
        "storage_test_swaps": null, 
        "storage_test_sys_node": null
    }, 
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Saturday 14 December 2024  17:55:06 -0500 (0:00:00.055)       0:00:53.826 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1", 
        "storage_test_fstab_expected_mount_options_matches": "1", 
        "storage_test_fstab_expected_mount_point_matches": "1", 
        "storage_test_fstab_id_matches": [
            "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269 "
        ], 
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ], 
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    }, 
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Saturday 14 December 2024  17:55:06 -0500 (0:00:00.102)       0:00:53.929 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Saturday 14 December 2024  17:55:06 -0500 (0:00:00.064)       0:00:53.993 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Saturday 14 December 2024  17:55:06 -0500 (0:00:00.062)       0:00:54.055 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Saturday 14 December 2024  17:55:06 -0500 (0:00:00.057)       0:00:54.113 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Saturday 14 December 2024  17:55:06 -0500 (0:00:00.071)       0:00:54.184 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null, 
        "storage_test_fstab_expected_mount_options_matches": null, 
        "storage_test_fstab_expected_mount_point_matches": null, 
        "storage_test_fstab_id_matches": null, 
        "storage_test_fstab_mount_options_matches": null, 
        "storage_test_fstab_mount_point_matches": null
    }, 
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Saturday 14 December 2024  17:55:06 -0500 (0:00:00.056)       0:00:54.241 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Saturday 14 December 2024  17:55:06 -0500 (0:00:00.067)       0:00:54.308 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Saturday 14 December 2024  17:55:06 -0500 (0:00:00.070)       0:00:54.379 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734216896.764867, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "ctime": 1734216896.764867, 
        "dev": 5, 
        "device_type": 2048, 
        "executable": false, 
        "exists": true, 
        "gid": 6, 
        "gr_name": "disk", 
        "inode": 28408, 
        "isblk": true, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": false, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/blockdevice", 
        "mode": "0660", 
        "mtime": 1734216896.764867, 
        "nlink": 1, 
        "path": "/dev/sda", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": null, 
        "wgrp": true, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Saturday 14 December 2024  17:55:07 -0500 (0:00:00.398)       0:00:54.777 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Saturday 14 December 2024  17:55:07 -0500 (0:00:00.066)       0:00:54.844 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Saturday 14 December 2024  17:55:07 -0500 (0:00:00.061)       0:00:54.905 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Saturday 14 December 2024  17:55:07 -0500 (0:00:00.067)       0:00:54.973 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "st_volume_type": "disk"
    }, 
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Saturday 14 December 2024  17:55:07 -0500 (0:00:00.066)       0:00:55.040 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Saturday 14 December 2024  17:55:07 -0500 (0:00:00.052)       0:00:55.093 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Saturday 14 December 2024  17:55:07 -0500 (0:00:00.063)       0:00:55.157 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734216896.879867, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "ctime": 1734216896.879867, 
        "dev": 5, 
        "device_type": 64768, 
        "executable": false, 
        "exists": true, 
        "gid": 6, 
        "gr_name": "disk", 
        "inode": 373405, 
        "isblk": true, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": false, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/symlink", 
        "mode": "0660", 
        "mtime": 1734216896.879867, 
        "nlink": 1, 
        "path": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": null, 
        "wgrp": true, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Saturday 14 December 2024  17:55:08 -0500 (0:00:00.392)       0:00:55.549 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}
lsrpackages: cryptsetup

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Saturday 14 December 2024  17:55:08 -0500 (0:00:00.733)       0:00:56.282 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cryptsetup", 
        "luksDump", 
        "/dev/sda"
    ], 
    "delta": "0:00:00.027533", 
    "end": "2024-12-14 17:55:09.314184", 
    "rc": 0, 
    "start": "2024-12-14 17:55:09.286651"
}

STDOUT:

LUKS header information for /dev/sda

Version:       	1
Cipher name:   	aes
Cipher mode:   	xts-plain64
Hash spec:     	sha256
Payload offset:	8192
MK bits:       	512
MK digest:     	57 88 7d ba 3a ed 1c b9 74 66 01 16 4f 25 c9 68 e6 54 08 68 
MK salt:       	1f 6d eb 0c 2b 56 3f 32 87 dc 22 48 47 82 37 ba 
               	87 c1 16 fc 47 e2 ba 41 15 f9 14 98 36 b5 8e 2d 
MK iterations: 	23206
UUID:          	b0c50623-e6eb-411b-b5d3-25165f6ee269

Key Slot 0: ENABLED
	Iterations:         	369736
	Salt:               	58 5b d5 2b f1 fd db 7e d5 f7 d4 d2 dd e7 1c 80 
	                      	8c e8 74 de 24 f5 7d b9 ac 5d 4e 06 b0 98 f5 98 
	Key material offset:	8
	AF stripes:            	4000
Key Slot 1: DISABLED
Key Slot 2: DISABLED
Key Slot 3: DISABLED
Key Slot 4: DISABLED
Key Slot 5: DISABLED
Key Slot 6: DISABLED
Key Slot 7: DISABLED

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Saturday 14 December 2024  17:55:09 -0500 (0:00:00.502)       0:00:56.785 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Saturday 14 December 2024  17:55:09 -0500 (0:00:00.068)       0:00:56.853 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Saturday 14 December 2024  17:55:09 -0500 (0:00:00.074)       0:00:56.927 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Saturday 14 December 2024  17:55:09 -0500 (0:00:00.074)       0:00:57.002 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Saturday 14 December 2024  17:55:09 -0500 (0:00:00.072)       0:00:57.074 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Saturday 14 December 2024  17:55:09 -0500 (0:00:00.082)       0:00:57.157 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Saturday 14 December 2024  17:55:09 -0500 (0:00:00.058)       0:00:57.215 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Saturday 14 December 2024  17:55:09 -0500 (0:00:00.067)       0:00:57.283 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [
            "luks-b0c50623-e6eb-411b-b5d3-25165f6ee269 /dev/sda -"
        ], 
        "_storage_test_expected_crypttab_entries": "1", 
        "_storage_test_expected_crypttab_key_file": "-"
    }, 
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Saturday 14 December 2024  17:55:09 -0500 (0:00:00.080)       0:00:57.363 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Saturday 14 December 2024  17:55:10 -0500 (0:00:00.082)       0:00:57.446 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Saturday 14 December 2024  17:55:10 -0500 (0:00:00.078)       0:00:57.525 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Saturday 14 December 2024  17:55:10 -0500 (0:00:00.079)       0:00:57.604 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Saturday 14 December 2024  17:55:10 -0500 (0:00:00.067)       0:00:57.671 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null, 
        "_storage_test_expected_crypttab_entries": null, 
        "_storage_test_expected_crypttab_key_file": null
    }, 
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Saturday 14 December 2024  17:55:10 -0500 (0:00:00.052)       0:00:57.724 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Saturday 14 December 2024  17:55:10 -0500 (0:00:00.060)       0:00:57.785 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Saturday 14 December 2024  17:55:10 -0500 (0:00:00.060)       0:00:57.845 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Saturday 14 December 2024  17:55:10 -0500 (0:00:00.050)       0:00:57.896 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Saturday 14 December 2024  17:55:10 -0500 (0:00:00.055)       0:00:57.951 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Saturday 14 December 2024  17:55:10 -0500 (0:00:00.055)       0:00:58.006 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Saturday 14 December 2024  17:55:10 -0500 (0:00:00.067)       0:00:58.073 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Saturday 14 December 2024  17:55:10 -0500 (0:00:00.054)       0:00:58.128 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Saturday 14 December 2024  17:55:10 -0500 (0:00:00.102)       0:00:58.230 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Saturday 14 December 2024  17:55:10 -0500 (0:00:00.056)       0:00:58.286 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Saturday 14 December 2024  17:55:10 -0500 (0:00:00.049)       0:00:58.336 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Saturday 14 December 2024  17:55:11 -0500 (0:00:00.050)       0:00:58.386 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Saturday 14 December 2024  17:55:11 -0500 (0:00:00.042)       0:00:58.429 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Saturday 14 December 2024  17:55:11 -0500 (0:00:00.039)       0:00:58.468 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Saturday 14 December 2024  17:55:11 -0500 (0:00:00.039)       0:00:58.508 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Saturday 14 December 2024  17:55:11 -0500 (0:00:00.047)       0:00:58.555 ***** 
skipping: [managed-node2] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Saturday 14 December 2024  17:55:11 -0500 (0:00:00.055)       0:00:58.611 ***** 
skipping: [managed-node2] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Saturday 14 December 2024  17:55:11 -0500 (0:00:00.057)       0:00:58.668 ***** 
skipping: [managed-node2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Saturday 14 December 2024  17:55:11 -0500 (0:00:00.054)       0:00:58.723 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Saturday 14 December 2024  17:55:11 -0500 (0:00:00.075)       0:00:58.798 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Saturday 14 December 2024  17:55:11 -0500 (0:00:00.070)       0:00:58.868 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Saturday 14 December 2024  17:55:11 -0500 (0:00:00.071)       0:00:58.940 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Saturday 14 December 2024  17:55:11 -0500 (0:00:00.079)       0:00:59.020 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Saturday 14 December 2024  17:55:11 -0500 (0:00:00.064)       0:00:59.084 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Saturday 14 December 2024  17:55:11 -0500 (0:00:00.052)       0:00:59.137 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Saturday 14 December 2024  17:55:11 -0500 (0:00:00.061)       0:00:59.199 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Saturday 14 December 2024  17:55:11 -0500 (0:00:00.056)       0:00:59.256 ***** 
skipping: [managed-node2] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Saturday 14 December 2024  17:55:11 -0500 (0:00:00.112)       0:00:59.369 ***** 
skipping: [managed-node2] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Saturday 14 December 2024  17:55:12 -0500 (0:00:00.070)       0:00:59.439 ***** 
skipping: [managed-node2] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Saturday 14 December 2024  17:55:12 -0500 (0:00:00.068)       0:00:59.507 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Saturday 14 December 2024  17:55:12 -0500 (0:00:00.056)       0:00:59.563 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Saturday 14 December 2024  17:55:12 -0500 (0:00:00.055)       0:00:59.618 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Saturday 14 December 2024  17:55:12 -0500 (0:00:00.064)       0:00:59.683 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Saturday 14 December 2024  17:55:12 -0500 (0:00:00.050)       0:00:59.734 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Saturday 14 December 2024  17:55:12 -0500 (0:00:00.073)       0:00:59.808 ***** 
ok: [managed-node2] => {
    "storage_test_actual_size": {
        "changed": false, 
        "skip_reason": "Conditional result was False", 
        "skipped": true
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Saturday 14 December 2024  17:55:12 -0500 (0:00:00.099)       0:00:59.907 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Saturday 14 December 2024  17:55:12 -0500 (0:00:00.108)       0:01:00.015 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Saturday 14 December 2024  17:55:12 -0500 (0:00:00.063)       0:01:00.079 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Saturday 14 December 2024  17:55:12 -0500 (0:00:00.081)       0:01:00.160 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Saturday 14 December 2024  17:55:12 -0500 (0:00:00.054)       0:01:00.215 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Saturday 14 December 2024  17:55:12 -0500 (0:00:00.055)       0:01:00.271 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Saturday 14 December 2024  17:55:12 -0500 (0:00:00.085)       0:01:00.357 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Saturday 14 December 2024  17:55:13 -0500 (0:00:00.053)       0:01:00.411 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Saturday 14 December 2024  17:55:13 -0500 (0:00:00.058)       0:01:00.470 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Saturday 14 December 2024  17:55:13 -0500 (0:00:00.056)       0:01:00.526 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    }, 
    "changed": false
}

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Saturday 14 December 2024  17:55:13 -0500 (0:00:00.054)       0:01:00.580 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null, 
        "storage_test_crypttab": null, 
        "storage_test_fstab": null
    }, 
    "changed": false
}

TASK [Create a file] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml:12
Saturday 14 December 2024  17:55:13 -0500 (0:00:00.054)       0:01:00.634 ***** 
changed: [managed-node2] => {
    "changed": true, 
    "dest": "/opt/test1/quux", 
    "gid": 0, 
    "group": "root", 
    "mode": "0644", 
    "owner": "root", 
    "secontext": "unconfined_u:object_r:unlabeled_t:s0", 
    "size": 0, 
    "state": "file", 
    "uid": 0
}

TASK [Test for correct handling of safe_mode] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:118
Saturday 14 December 2024  17:55:14 -0500 (0:00:00.855)       0:01:01.489 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node2

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Saturday 14 December 2024  17:55:14 -0500 (0:00:00.131)       0:01:01.621 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_pools_global": [], 
        "storage_safe_mode_global": true, 
        "storage_volumes_global": []
    }, 
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Saturday 14 December 2024  17:55:14 -0500 (0:00:00.071)       0:01:01.693 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  17:55:14 -0500 (0:00:00.175)       0:01:01.868 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  17:55:14 -0500 (0:00:00.111)       0:01:01.979 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  17:55:14 -0500 (0:00:00.080)       0:01:02.059 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  17:55:14 -0500 (0:00:00.132)       0:01:02.192 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  17:55:14 -0500 (0:00:00.069)       0:01:02.261 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  17:55:14 -0500 (0:00:00.052)       0:01:02.314 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  17:55:14 -0500 (0:00:00.066)       0:01:02.380 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  17:55:15 -0500 (0:00:00.072)       0:01:02.453 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  17:55:15 -0500 (0:00:00.171)       0:01:02.624 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  17:55:16 -0500 (0:00:01.277)       0:01:03.902 ***** 
ok: [managed-node2] => {
    "storage_pools": []
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  17:55:16 -0500 (0:00:00.056)       0:01:03.958 ***** 
ok: [managed-node2] => {
    "storage_volumes": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_password": "yabbadabbadoo", 
            "mount_point": "/opt/test1", 
            "name": "foo", 
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  17:55:16 -0500 (0:00:00.069)       0:01:04.027 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  17:55:20 -0500 (0:00:04.025)       0:01:08.053 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  17:55:20 -0500 (0:00:00.120)       0:01:08.173 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  17:55:20 -0500 (0:00:00.062)       0:01:08.236 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  17:55:20 -0500 (0:00:00.057)       0:01:08.294 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  17:55:20 -0500 (0:00:00.052)       0:01:08.346 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: kpartx

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  17:55:21 -0500 (0:00:00.895)       0:01:09.242 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  17:55:23 -0500 (0:00:01.323)       0:01:10.565 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  17:55:23 -0500 (0:00:00.092)       0:01:10.658 ***** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  17:55:23 -0500 (0:00:00.068)       0:01:10.726 ***** 
fatal: [managed-node2]: FAILED! => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [], 
    "pools": [], 
    "volumes": []
}

MSG:

cannot remove existing formatting on device 'luks-b0c50623-e6eb-411b-b5d3-25165f6ee269' in safe mode due to encryption removal

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Saturday 14 December 2024  17:55:27 -0500 (0:00:04.056)       0:01:14.782 ***** 
fatal: [managed-node2]: FAILED! => {
    "changed": false
}

MSG:

{u'_ansible_no_log': False, u'crypts': [], u'pools': [], u'leaves': [], u'changed': False, u'actions': [], u'failed': True, u'volumes': [], u'invocation': {u'module_args': {u'packages_only': False, u'disklabel_type': None, u'diskvolume_mkfs_option_map': {u'ext4': u'-F', u'ext3': u'-F', u'ext2': u'-F'}, u'safe_mode': True, u'pools': [], u'volumes': [{u'raid_metadata_version': None, u'raid_level': None, u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': 10733223936, u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'encryption': False, u'mount_device_identifier': u'uuid', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'fs_overwrite_existing': True, u'encryption_key_size': 0, u'deduplication': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': u'luks1', u'raid_stripe_size': None, u'cache_size': 0, u'mount_user': None, u'raid_spare_count': None, u'cache_mode': None, u'name': u'foo', u'mount_group': None, u'type': u'disk', u'disks': [u'sda'], u'cached': False, u'thin_pool_size': None, u'thin': None, u'mount_check': 0, u'mount_passno': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}], u'pool_defaults': {u'raid_metadata_version': None, u'encryption_cipher': None, u'encryption_key': None, u'encryption_luks_version': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_password': None, u'encryption': False, u'disks': [], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [], u'shared': False, u'type': u'lvm', u'encryption_key_size': None, u'raid_chunk_size': None}, u'volume_defaults': {u'raid_metadata_version': None, u'raid_level': None, u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': 0, u'mount_point': u'', u'compression': None, u'encryption_password': None, u'encryption': False, u'mount_device_identifier': u'uuid', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'thin_pool_name': None, u'type': u'lvm', u'encryption_key_size': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'cache_size': 0, u'raid_spare_count': None, u'cache_mode': None, u'deduplication': None, u'cached': False, u'fs_overwrite_existing': True, u'disks': [], u'thin_pool_size': None, u'thin': None, u'mount_check': 0, u'mount_passno': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}, u'use_partitions': None}}, u'mounts': [], u'packages': [], u'msg': u"cannot remove existing formatting on device 'luks-b0c50623-e6eb-411b-b5d3-25165f6ee269' in safe mode due to encryption removal"}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  17:55:27 -0500 (0:00:00.075)       0:01:14.858 ***** 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Saturday 14 December 2024  17:55:27 -0500 (0:00:00.050)       0:01:14.908 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Saturday 14 December 2024  17:55:27 -0500 (0:00:00.060)       0:01:14.969 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Saturday 14 December 2024  17:55:27 -0500 (0:00:00.090)       0:01:15.059 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Stat the file] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:11
Saturday 14 December 2024  17:55:27 -0500 (0:00:00.057)       0:01:15.117 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734216913.988859, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", 
        "ctime": 1734216913.988859, 
        "dev": 64768, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 67, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/x-empty", 
        "mode": "0644", 
        "mtime": 1734216913.988859, 
        "nlink": 1, 
        "path": "/opt/test1/quux", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": "18446744072002628453", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Assert file presence] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:16
Saturday 14 December 2024  17:55:28 -0500 (0:00:00.543)       0:01:15.660 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Remove the encryption layer] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:138
Saturday 14 December 2024  17:55:28 -0500 (0:00:00.067)       0:01:15.728 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  17:55:28 -0500 (0:00:00.182)       0:01:15.911 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  17:55:28 -0500 (0:00:00.107)       0:01:16.018 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  17:55:28 -0500 (0:00:00.082)       0:01:16.103 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  17:55:28 -0500 (0:00:00.144)       0:01:16.247 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  17:55:28 -0500 (0:00:00.054)       0:01:16.302 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  17:55:28 -0500 (0:00:00.062)       0:01:16.365 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  17:55:29 -0500 (0:00:00.052)       0:01:16.417 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  17:55:29 -0500 (0:00:00.108)       0:01:16.526 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  17:55:29 -0500 (0:00:00.192)       0:01:16.718 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  17:55:30 -0500 (0:00:01.295)       0:01:18.014 ***** 
ok: [managed-node2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  17:55:30 -0500 (0:00:00.058)       0:01:18.073 ***** 
ok: [managed-node2] => {
    "storage_volumes": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_password": "yabbadabbadoo", 
            "mount_point": "/opt/test1", 
            "name": "foo", 
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  17:55:30 -0500 (0:00:00.076)       0:01:18.150 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  17:55:34 -0500 (0:00:03.780)       0:01:21.931 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  17:55:34 -0500 (0:00:00.140)       0:01:22.071 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  17:55:34 -0500 (0:00:00.050)       0:01:22.122 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  17:55:34 -0500 (0:00:00.054)       0:01:22.177 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  17:55:34 -0500 (0:00:00.057)       0:01:22.234 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: kpartx

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  17:55:35 -0500 (0:00:00.924)       0:01:23.158 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  17:55:37 -0500 (0:00:01.357)       0:01:24.516 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  17:55:37 -0500 (0:00:00.124)       0:01:24.641 ***** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  17:55:37 -0500 (0:00:00.071)       0:01:24.712 ***** 
changed: [managed-node2] => {
    "actions": [
        {
            "action": "destroy format", 
            "device": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
            "fs_type": "xfs"
        }, 
        {
            "action": "destroy device", 
            "device": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
            "fs_type": null
        }, 
        {
            "action": "destroy format", 
            "device": "/dev/sda", 
            "fs_type": "luks"
        }, 
        {
            "action": "create format", 
            "device": "/dev/sda", 
            "fs_type": "xfs"
        }
    ], 
    "changed": true, 
    "crypts": [
        {
            "backing_device": "/dev/sda", 
            "name": "luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
            "password": "-", 
            "state": "absent"
        }
    ], 
    "leaves": [
        "/dev/sda", 
        "/dev/sdb", 
        "/dev/sdc", 
        "/dev/sdd", 
        "/dev/sde", 
        "/dev/sdf", 
        "/dev/sdg", 
        "/dev/sdh", 
        "/dev/sdi", 
        "/dev/xvda1"
    ], 
    "mounts": [
        {
            "fstype": "xfs", 
            "path": "/opt/test1", 
            "src": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
            "state": "absent"
        }, 
        {
            "dump": 0, 
            "fstype": "xfs", 
            "group": null, 
            "mode": null, 
            "opts": "defaults", 
            "owner": null, 
            "passno": 0, 
            "path": "/opt/test1", 
            "src": "UUID=6b177eed-d54c-42ab-b8d1-951cc9acb8aa", 
            "state": "mounted"
        }
    ], 
    "packages": [
        "xfsprogs", 
        "e2fsprogs"
    ], 
    "pools": [], 
    "volumes": [
        {
            "_device": "/dev/sda", 
            "_kernel_device": "/dev/sda", 
            "_mount_id": "UUID=6b177eed-d54c-42ab-b8d1-951cc9acb8aa", 
            "_raw_device": "/dev/sda", 
            "_raw_kernel_device": "/dev/sda", 
            "cache_devices": [], 
            "cache_mode": null, 
            "cache_size": 0, 
            "cached": false, 
            "compression": null, 
            "deduplication": null, 
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_cipher": null, 
            "encryption_key": null, 
            "encryption_key_size": 0, 
            "encryption_luks_version": "luks1", 
            "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
            "fs_create_options": "", 
            "fs_label": "", 
            "fs_overwrite_existing": true, 
            "fs_type": "xfs", 
            "mount_check": 0, 
            "mount_device_identifier": "uuid", 
            "mount_group": null, 
            "mount_mode": null, 
            "mount_options": "defaults", 
            "mount_passno": 0, 
            "mount_point": "/opt/test1", 
            "mount_user": null, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "raid_stripe_size": null, 
            "size": 10733223936, 
            "state": "present", 
            "thin": null, 
            "thin_pool_name": null, 
            "thin_pool_size": null, 
            "type": "disk", 
            "vdo_pool_size": null
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Saturday 14 December 2024  17:55:41 -0500 (0:00:04.360)       0:01:29.072 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Saturday 14 December 2024  17:55:41 -0500 (0:00:00.056)       0:01:29.129 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734216900.3988655, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "4ee9955ea93e985e966e94859f997cda72bd69df", 
        "ctime": 1734216900.3968654, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 263588, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0644", 
        "mtime": 1734216900.3968654, 
        "nlink": 1, 
        "path": "/etc/fstab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 1311, 
        "uid": 0, 
        "version": "606402765", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Saturday 14 December 2024  17:55:42 -0500 (0:00:00.448)       0:01:29.578 ***** 
ok: [managed-node2] => {
    "backup": "", 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  17:55:42 -0500 (0:00:00.460)       0:01:30.038 ***** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Saturday 14 December 2024  17:55:42 -0500 (0:00:00.052)       0:01:30.091 ***** 
ok: [managed-node2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format", 
                "device": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
                "fs_type": "xfs"
            }, 
            {
                "action": "destroy device", 
                "device": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
                "fs_type": null
            }, 
            {
                "action": "destroy format", 
                "device": "/dev/sda", 
                "fs_type": "luks"
            }, 
            {
                "action": "create format", 
                "device": "/dev/sda", 
                "fs_type": "xfs"
            }
        ], 
        "changed": true, 
        "crypts": [
            {
                "backing_device": "/dev/sda", 
                "name": "luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
                "password": "-", 
                "state": "absent"
            }
        ], 
        "failed": false, 
        "leaves": [
            "/dev/sda", 
            "/dev/sdb", 
            "/dev/sdc", 
            "/dev/sdd", 
            "/dev/sde", 
            "/dev/sdf", 
            "/dev/sdg", 
            "/dev/sdh", 
            "/dev/sdi", 
            "/dev/xvda1"
        ], 
        "mounts": [
            {
                "fstype": "xfs", 
                "path": "/opt/test1", 
                "src": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
                "state": "absent"
            }, 
            {
                "dump": 0, 
                "fstype": "xfs", 
                "group": null, 
                "mode": null, 
                "opts": "defaults", 
                "owner": null, 
                "passno": 0, 
                "path": "/opt/test1", 
                "src": "UUID=6b177eed-d54c-42ab-b8d1-951cc9acb8aa", 
                "state": "mounted"
            }
        ], 
        "packages": [
            "xfsprogs", 
            "e2fsprogs"
        ], 
        "pools": [], 
        "volumes": [
            {
                "_device": "/dev/sda", 
                "_kernel_device": "/dev/sda", 
                "_mount_id": "UUID=6b177eed-d54c-42ab-b8d1-951cc9acb8aa", 
                "_raw_device": "/dev/sda", 
                "_raw_kernel_device": "/dev/sda", 
                "cache_devices": [], 
                "cache_mode": null, 
                "cache_size": 0, 
                "cached": false, 
                "compression": null, 
                "deduplication": null, 
                "disks": [
                    "sda"
                ], 
                "encryption": false, 
                "encryption_cipher": null, 
                "encryption_key": null, 
                "encryption_key_size": 0, 
                "encryption_luks_version": "luks1", 
                "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                "fs_create_options": "", 
                "fs_label": "", 
                "fs_overwrite_existing": true, 
                "fs_type": "xfs", 
                "mount_check": 0, 
                "mount_device_identifier": "uuid", 
                "mount_group": null, 
                "mount_mode": null, 
                "mount_options": "defaults", 
                "mount_passno": 0, 
                "mount_point": "/opt/test1", 
                "mount_user": null, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "raid_stripe_size": null, 
                "size": 10733223936, 
                "state": "present", 
                "thin": null, 
                "thin_pool_name": null, 
                "thin_pool_size": null, 
                "type": "disk", 
                "vdo_pool_size": null
            }
        ]
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Saturday 14 December 2024  17:55:42 -0500 (0:00:00.088)       0:01:30.179 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Saturday 14 December 2024  17:55:42 -0500 (0:00:00.089)       0:01:30.269 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": [
            {
                "_device": "/dev/sda", 
                "_kernel_device": "/dev/sda", 
                "_mount_id": "UUID=6b177eed-d54c-42ab-b8d1-951cc9acb8aa", 
                "_raw_device": "/dev/sda", 
                "_raw_kernel_device": "/dev/sda", 
                "cache_devices": [], 
                "cache_mode": null, 
                "cache_size": 0, 
                "cached": false, 
                "compression": null, 
                "deduplication": null, 
                "disks": [
                    "sda"
                ], 
                "encryption": false, 
                "encryption_cipher": null, 
                "encryption_key": null, 
                "encryption_key_size": 0, 
                "encryption_luks_version": "luks1", 
                "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                "fs_create_options": "", 
                "fs_label": "", 
                "fs_overwrite_existing": true, 
                "fs_type": "xfs", 
                "mount_check": 0, 
                "mount_device_identifier": "uuid", 
                "mount_group": null, 
                "mount_mode": null, 
                "mount_options": "defaults", 
                "mount_passno": 0, 
                "mount_point": "/opt/test1", 
                "mount_user": null, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "raid_stripe_size": null, 
                "size": 10733223936, 
                "state": "present", 
                "thin": null, 
                "thin_pool_name": null, 
                "thin_pool_size": null, 
                "type": "disk", 
                "vdo_pool_size": null
            }
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Saturday 14 December 2024  17:55:42 -0500 (0:00:00.080)       0:01:30.349 ***** 
changed: [managed-node2] => (item={u'src': u'/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269', u'state': u'absent', u'fstype': u'xfs', u'path': u'/opt/test1'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": true, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "fstype": "xfs", 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
        "state": "absent"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "/dev/mapper/luks-b0c50623-e6eb-411b-b5d3-25165f6ee269"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Saturday 14 December 2024  17:55:43 -0500 (0:00:00.581)       0:01:30.930 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Saturday 14 December 2024  17:55:44 -0500 (0:00:00.600)       0:01:31.531 ***** 
changed: [managed-node2] => (item={u'src': u'UUID=6b177eed-d54c-42ab-b8d1-951cc9acb8aa', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": true, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "UUID=6b177eed-d54c-42ab-b8d1-951cc9acb8aa", 
        "state": "mounted"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "UUID=6b177eed-d54c-42ab-b8d1-951cc9acb8aa"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Saturday 14 December 2024  17:55:44 -0500 (0:00:00.523)       0:01:32.054 ***** 
skipping: [managed-node2] => (item={u'src': u'UUID=6b177eed-d54c-42ab-b8d1-951cc9acb8aa', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'})  => {
    "ansible_loop_var": "mount_info", 
    "changed": false, 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "UUID=6b177eed-d54c-42ab-b8d1-951cc9acb8aa", 
        "state": "mounted"
    }, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Saturday 14 December 2024  17:55:44 -0500 (0:00:00.082)       0:01:32.137 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Saturday 14 December 2024  17:55:45 -0500 (0:00:00.610)       0:01:32.747 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734216905.266863, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "196b6bcb21d17e3b8b4558c8e00a5ccd5b6f65a6", 
        "ctime": 1734216901.9398646, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 917511, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0600", 
        "mtime": 1734216901.9388647, 
        "nlink": 1, 
        "path": "/etc/crypttab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": false, 
        "roth": false, 
        "rusr": true, 
        "size": 53, 
        "uid": 0, 
        "version": "606408562", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Saturday 14 December 2024  17:55:45 -0500 (0:00:00.490)       0:01:33.238 ***** 
changed: [managed-node2] => (item={u'state': u'absent', u'password': u'-', u'name': u'luks-b0c50623-e6eb-411b-b5d3-25165f6ee269', u'backing_device': u'/dev/sda'}) => {
    "ansible_loop_var": "entry", 
    "backup": "", 
    "changed": true, 
    "entry": {
        "backing_device": "/dev/sda", 
        "name": "luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
        "password": "-", 
        "state": "absent"
    }, 
    "found": 1
}

MSG:

1 line(s) removed

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Saturday 14 December 2024  17:55:46 -0500 (0:00:00.450)       0:01:33.688 ***** 
ok: [managed-node2]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:151
Saturday 14 December 2024  17:55:47 -0500 (0:00:00.884)       0:01:34.573 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node2

TASK [Print out pool information] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Saturday 14 December 2024  17:55:47 -0500 (0:00:00.198)       0:01:34.771 ***** 
skipping: [managed-node2] => {}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Saturday 14 December 2024  17:55:47 -0500 (0:00:00.089)       0:01:34.861 ***** 
ok: [managed-node2] => {
    "_storage_volumes_list": [
        {
            "_device": "/dev/sda", 
            "_kernel_device": "/dev/sda", 
            "_mount_id": "UUID=6b177eed-d54c-42ab-b8d1-951cc9acb8aa", 
            "_raw_device": "/dev/sda", 
            "_raw_kernel_device": "/dev/sda", 
            "cache_devices": [], 
            "cache_mode": null, 
            "cache_size": 0, 
            "cached": false, 
            "compression": null, 
            "deduplication": null, 
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_cipher": null, 
            "encryption_key": null, 
            "encryption_key_size": 0, 
            "encryption_luks_version": "luks1", 
            "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
            "fs_create_options": "", 
            "fs_label": "", 
            "fs_overwrite_existing": true, 
            "fs_type": "xfs", 
            "mount_check": 0, 
            "mount_device_identifier": "uuid", 
            "mount_group": null, 
            "mount_mode": null, 
            "mount_options": "defaults", 
            "mount_passno": 0, 
            "mount_point": "/opt/test1", 
            "mount_user": null, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "raid_stripe_size": null, 
            "size": 10733223936, 
            "state": "present", 
            "thin": null, 
            "thin_pool_name": null, 
            "thin_pool_size": null, 
            "type": "disk", 
            "vdo_pool_size": null
        }
    ]
}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Saturday 14 December 2024  17:55:47 -0500 (0:00:00.072)       0:01:34.933 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "info": {
        "/dev/sda": {
            "fstype": "xfs", 
            "label": "", 
            "mountpoint": "/opt/test1", 
            "name": "/dev/sda", 
            "size": "10G", 
            "type": "disk", 
            "uuid": "6b177eed-d54c-42ab-b8d1-951cc9acb8aa"
        }, 
        "/dev/sdb": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdb", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdc": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdc", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdd": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdd", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sde": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sde", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdf": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdf", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdg": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdg", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdh": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdh", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdi": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdi", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/xvda", 
            "size": "250G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda1": {
            "fstype": "ext4", 
            "label": "", 
            "mountpoint": "/", 
            "name": "/dev/xvda1", 
            "size": "250G", 
            "type": "partition", 
            "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Saturday 14 December 2024  17:55:48 -0500 (0:00:00.623)       0:01:35.556 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/fstab"
    ], 
    "delta": "0:00:00.002948", 
    "end": "2024-12-14 17:55:48.504396", 
    "rc": 0, 
    "start": "2024-12-14 17:55:48.501448"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Thu Jun 20 10:23:46 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 /                       ext4    defaults        1 1
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
UUID=6b177eed-d54c-42ab-b8d1-951cc9acb8aa /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Saturday 14 December 2024  17:55:48 -0500 (0:00:00.411)       0:01:35.967 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/crypttab"
    ], 
    "delta": "0:00:00.002940", 
    "end": "2024-12-14 17:55:48.989802", 
    "failed_when_result": false, 
    "rc": 0, 
    "start": "2024-12-14 17:55:48.986862"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Saturday 14 December 2024  17:55:49 -0500 (0:00:00.517)       0:01:36.485 ***** 

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Saturday 14 December 2024  17:55:49 -0500 (0:00:00.075)       0:01:36.561 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node2

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Saturday 14 December 2024  17:55:49 -0500 (0:00:00.117)       0:01:36.679 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true, 
        "_storage_volume_tests": [
            "mount", 
            "fstab", 
            "fs", 
            "device", 
            "encryption", 
            "md", 
            "size", 
            "cache"
        ]
    }, 
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Saturday 14 December 2024  17:55:49 -0500 (0:00:00.063)       0:01:36.743 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node2

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Saturday 14 December 2024  17:55:49 -0500 (0:00:00.310)       0:01:37.053 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/sda"
    }, 
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Saturday 14 December 2024  17:55:49 -0500 (0:00:00.117)       0:01:37.170 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1", 
        "storage_test_swap_expected_matches": "0"
    }, 
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Saturday 14 December 2024  17:55:49 -0500 (0:00:00.083)       0:01:37.254 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Saturday 14 December 2024  17:55:49 -0500 (0:00:00.102)       0:01:37.356 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Saturday 14 December 2024  17:55:50 -0500 (0:00:00.149)       0:01:37.506 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Saturday 14 December 2024  17:55:50 -0500 (0:00:00.109)       0:01:37.615 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Saturday 14 December 2024  17:55:50 -0500 (0:00:00.157)       0:01:37.773 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Saturday 14 December 2024  17:55:50 -0500 (0:00:00.116)       0:01:37.889 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Saturday 14 December 2024  17:55:50 -0500 (0:00:00.094)       0:01:37.983 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Saturday 14 December 2024  17:55:50 -0500 (0:00:00.063)       0:01:38.046 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Saturday 14 December 2024  17:55:50 -0500 (0:00:00.133)       0:01:38.179 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null, 
        "storage_test_mount_expected_mount_point": null, 
        "storage_test_swap_expected_matches": null, 
        "storage_test_swaps": null, 
        "storage_test_sys_node": null
    }, 
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Saturday 14 December 2024  17:55:50 -0500 (0:00:00.067)       0:01:38.247 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1", 
        "storage_test_fstab_expected_mount_options_matches": "1", 
        "storage_test_fstab_expected_mount_point_matches": "1", 
        "storage_test_fstab_id_matches": [
            "UUID=6b177eed-d54c-42ab-b8d1-951cc9acb8aa "
        ], 
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ], 
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    }, 
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Saturday 14 December 2024  17:55:50 -0500 (0:00:00.114)       0:01:38.361 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Saturday 14 December 2024  17:55:51 -0500 (0:00:00.065)       0:01:38.426 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Saturday 14 December 2024  17:55:51 -0500 (0:00:00.065)       0:01:38.492 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Saturday 14 December 2024  17:55:51 -0500 (0:00:00.052)       0:01:38.545 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Saturday 14 December 2024  17:55:51 -0500 (0:00:00.124)       0:01:38.669 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null, 
        "storage_test_fstab_expected_mount_options_matches": null, 
        "storage_test_fstab_expected_mount_point_matches": null, 
        "storage_test_fstab_id_matches": null, 
        "storage_test_fstab_mount_options_matches": null, 
        "storage_test_fstab_mount_point_matches": null
    }, 
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Saturday 14 December 2024  17:55:51 -0500 (0:00:00.119)       0:01:38.789 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Saturday 14 December 2024  17:55:51 -0500 (0:00:00.115)       0:01:38.904 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Saturday 14 December 2024  17:55:51 -0500 (0:00:00.144)       0:01:39.048 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734216941.532846, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "ctime": 1734216941.532846, 
        "dev": 5, 
        "device_type": 2048, 
        "executable": false, 
        "exists": true, 
        "gid": 6, 
        "gr_name": "disk", 
        "inode": 28408, 
        "isblk": true, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": false, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/blockdevice", 
        "mode": "0660", 
        "mtime": 1734216941.532846, 
        "nlink": 1, 
        "path": "/dev/sda", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": null, 
        "wgrp": true, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Saturday 14 December 2024  17:55:52 -0500 (0:00:00.463)       0:01:39.511 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Saturday 14 December 2024  17:55:52 -0500 (0:00:00.096)       0:01:39.609 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Saturday 14 December 2024  17:55:52 -0500 (0:00:00.074)       0:01:39.683 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Saturday 14 December 2024  17:55:52 -0500 (0:00:00.063)       0:01:39.746 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "st_volume_type": "disk"
    }, 
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Saturday 14 December 2024  17:55:52 -0500 (0:00:00.068)       0:01:39.814 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Saturday 14 December 2024  17:55:52 -0500 (0:00:00.055)       0:01:39.870 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Saturday 14 December 2024  17:55:52 -0500 (0:00:00.068)       0:01:39.938 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Saturday 14 December 2024  17:55:52 -0500 (0:00:00.054)       0:01:39.993 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}
lsrpackages: cryptsetup

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Saturday 14 December 2024  17:55:53 -0500 (0:00:00.679)       0:01:40.673 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Saturday 14 December 2024  17:55:53 -0500 (0:00:00.090)       0:01:40.763 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Saturday 14 December 2024  17:55:53 -0500 (0:00:00.092)       0:01:40.856 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Saturday 14 December 2024  17:55:53 -0500 (0:00:00.164)       0:01:41.020 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Saturday 14 December 2024  17:55:53 -0500 (0:00:00.057)       0:01:41.078 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Saturday 14 December 2024  17:55:53 -0500 (0:00:00.054)       0:01:41.133 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Saturday 14 December 2024  17:55:53 -0500 (0:00:00.063)       0:01:41.197 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Saturday 14 December 2024  17:55:53 -0500 (0:00:00.104)       0:01:41.301 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Saturday 14 December 2024  17:55:54 -0500 (0:00:00.104)       0:01:41.406 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [], 
        "_storage_test_expected_crypttab_entries": "0", 
        "_storage_test_expected_crypttab_key_file": "-"
    }, 
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Saturday 14 December 2024  17:55:54 -0500 (0:00:00.136)       0:01:41.542 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Saturday 14 December 2024  17:55:54 -0500 (0:00:00.148)       0:01:41.691 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Saturday 14 December 2024  17:55:54 -0500 (0:00:00.067)       0:01:41.758 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Saturday 14 December 2024  17:55:54 -0500 (0:00:00.057)       0:01:41.816 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Saturday 14 December 2024  17:55:54 -0500 (0:00:00.056)       0:01:41.872 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null, 
        "_storage_test_expected_crypttab_entries": null, 
        "_storage_test_expected_crypttab_key_file": null
    }, 
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Saturday 14 December 2024  17:55:54 -0500 (0:00:00.056)       0:01:41.928 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Saturday 14 December 2024  17:55:54 -0500 (0:00:00.055)       0:01:41.984 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Saturday 14 December 2024  17:55:54 -0500 (0:00:00.062)       0:01:42.047 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Saturday 14 December 2024  17:55:54 -0500 (0:00:00.056)       0:01:42.103 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Saturday 14 December 2024  17:55:54 -0500 (0:00:00.072)       0:01:42.175 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Saturday 14 December 2024  17:55:54 -0500 (0:00:00.055)       0:01:42.231 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Saturday 14 December 2024  17:55:54 -0500 (0:00:00.078)       0:01:42.310 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Saturday 14 December 2024  17:55:55 -0500 (0:00:00.116)       0:01:42.426 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Saturday 14 December 2024  17:55:55 -0500 (0:00:00.091)       0:01:42.517 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Saturday 14 December 2024  17:55:55 -0500 (0:00:00.066)       0:01:42.584 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Saturday 14 December 2024  17:55:55 -0500 (0:00:00.117)       0:01:42.702 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Saturday 14 December 2024  17:55:55 -0500 (0:00:00.059)       0:01:42.762 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Saturday 14 December 2024  17:55:55 -0500 (0:00:00.061)       0:01:42.823 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Saturday 14 December 2024  17:55:55 -0500 (0:00:00.071)       0:01:42.895 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Saturday 14 December 2024  17:55:55 -0500 (0:00:00.069)       0:01:42.964 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Saturday 14 December 2024  17:55:55 -0500 (0:00:00.124)       0:01:43.089 ***** 
skipping: [managed-node2] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Saturday 14 December 2024  17:55:55 -0500 (0:00:00.075)       0:01:43.164 ***** 
skipping: [managed-node2] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Saturday 14 December 2024  17:55:55 -0500 (0:00:00.056)       0:01:43.221 ***** 
skipping: [managed-node2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Saturday 14 December 2024  17:55:55 -0500 (0:00:00.062)       0:01:43.283 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Saturday 14 December 2024  17:55:55 -0500 (0:00:00.061)       0:01:43.345 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Saturday 14 December 2024  17:55:56 -0500 (0:00:00.054)       0:01:43.400 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Saturday 14 December 2024  17:55:56 -0500 (0:00:00.070)       0:01:43.471 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Saturday 14 December 2024  17:55:56 -0500 (0:00:00.068)       0:01:43.540 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Saturday 14 December 2024  17:55:56 -0500 (0:00:00.079)       0:01:43.619 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Saturday 14 December 2024  17:55:56 -0500 (0:00:00.126)       0:01:43.746 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Saturday 14 December 2024  17:55:56 -0500 (0:00:00.056)       0:01:43.802 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Saturday 14 December 2024  17:55:56 -0500 (0:00:00.059)       0:01:43.862 ***** 
skipping: [managed-node2] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Saturday 14 December 2024  17:55:56 -0500 (0:00:00.070)       0:01:43.933 ***** 
skipping: [managed-node2] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Saturday 14 December 2024  17:55:56 -0500 (0:00:00.079)       0:01:44.013 ***** 
skipping: [managed-node2] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Saturday 14 December 2024  17:55:56 -0500 (0:00:00.055)       0:01:44.069 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Saturday 14 December 2024  17:55:56 -0500 (0:00:00.055)       0:01:44.124 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Saturday 14 December 2024  17:55:56 -0500 (0:00:00.054)       0:01:44.178 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Saturday 14 December 2024  17:55:56 -0500 (0:00:00.052)       0:01:44.231 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Saturday 14 December 2024  17:55:56 -0500 (0:00:00.058)       0:01:44.290 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Saturday 14 December 2024  17:55:56 -0500 (0:00:00.060)       0:01:44.351 ***** 
ok: [managed-node2] => {
    "storage_test_actual_size": {
        "changed": false, 
        "skip_reason": "Conditional result was False", 
        "skipped": true
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Saturday 14 December 2024  17:55:57 -0500 (0:00:00.063)       0:01:44.414 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Saturday 14 December 2024  17:55:57 -0500 (0:00:00.054)       0:01:44.469 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Saturday 14 December 2024  17:55:57 -0500 (0:00:00.061)       0:01:44.531 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Saturday 14 December 2024  17:55:57 -0500 (0:00:00.051)       0:01:44.582 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Saturday 14 December 2024  17:55:57 -0500 (0:00:00.070)       0:01:44.653 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Saturday 14 December 2024  17:55:57 -0500 (0:00:00.075)       0:01:44.728 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Saturday 14 December 2024  17:55:57 -0500 (0:00:00.056)       0:01:44.785 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Saturday 14 December 2024  17:55:57 -0500 (0:00:00.061)       0:01:44.847 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Saturday 14 December 2024  17:55:57 -0500 (0:00:00.060)       0:01:44.907 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Saturday 14 December 2024  17:55:57 -0500 (0:00:00.083)       0:01:44.990 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    }, 
    "changed": false
}

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Saturday 14 December 2024  17:55:57 -0500 (0:00:00.054)       0:01:45.045 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null, 
        "storage_test_crypttab": null, 
        "storage_test_fstab": null
    }, 
    "changed": false
}

TASK [Create a file] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml:12
Saturday 14 December 2024  17:55:57 -0500 (0:00:00.052)       0:01:45.098 ***** 
changed: [managed-node2] => {
    "changed": true, 
    "dest": "/opt/test1/quux", 
    "gid": 0, 
    "group": "root", 
    "mode": "0644", 
    "owner": "root", 
    "secontext": "unconfined_u:object_r:unlabeled_t:s0", 
    "size": 0, 
    "state": "file", 
    "uid": 0
}

TASK [Test for correct handling of safe_mode] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:157
Saturday 14 December 2024  17:55:58 -0500 (0:00:00.347)       0:01:45.445 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node2

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Saturday 14 December 2024  17:55:58 -0500 (0:00:00.087)       0:01:45.533 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_pools_global": [], 
        "storage_safe_mode_global": true, 
        "storage_volumes_global": []
    }, 
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Saturday 14 December 2024  17:55:58 -0500 (0:00:00.065)       0:01:45.599 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  17:55:58 -0500 (0:00:00.133)       0:01:45.732 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  17:55:58 -0500 (0:00:00.113)       0:01:45.846 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  17:55:58 -0500 (0:00:00.099)       0:01:45.945 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  17:55:58 -0500 (0:00:00.174)       0:01:46.120 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  17:55:58 -0500 (0:00:00.061)       0:01:46.182 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  17:55:58 -0500 (0:00:00.062)       0:01:46.244 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  17:55:58 -0500 (0:00:00.110)       0:01:46.355 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  17:55:59 -0500 (0:00:00.106)       0:01:46.461 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  17:55:59 -0500 (0:00:00.150)       0:01:46.612 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  17:56:00 -0500 (0:00:01.399)       0:01:48.011 ***** 
ok: [managed-node2] => {
    "storage_pools": []
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  17:56:00 -0500 (0:00:00.071)       0:01:48.082 ***** 
ok: [managed-node2] => {
    "storage_volumes": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": true, 
            "encryption_password": "yabbadabbadoo", 
            "mount_point": "/opt/test1", 
            "name": "foo", 
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  17:56:00 -0500 (0:00:00.126)       0:01:48.209 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [
        "cryptsetup"
    ], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  17:56:04 -0500 (0:00:04.168)       0:01:52.377 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  17:56:05 -0500 (0:00:00.109)       0:01:52.487 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  17:56:05 -0500 (0:00:00.053)       0:01:52.541 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  17:56:05 -0500 (0:00:00.046)       0:01:52.587 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  17:56:05 -0500 (0:00:00.041)       0:01:52.628 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed", 
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: cryptsetup kpartx

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  17:56:05 -0500 (0:00:00.730)       0:01:53.359 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-cryptsetup@luks\\x2db0c50623\\x2de6eb\\x2d411b\\x2db5d3\\x2d25165f6ee269.service": {
                "name": "systemd-cryptsetup@luks\\x2db0c50623\\x2de6eb\\x2d411b\\x2db5d3\\x2d25165f6ee269.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  17:56:07 -0500 (0:00:01.053)       0:01:54.412 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": [
            "systemd-cryptsetup@luks\\x2db0c50623\\x2de6eb\\x2d411b\\x2db5d3\\x2d25165f6ee269.service"
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  17:56:07 -0500 (0:00:00.061)       0:01:54.474 ***** 
changed: [managed-node2] => (item=systemd-cryptsetup@luks\x2db0c50623\x2de6eb\x2d411b\x2db5d3\x2d25165f6ee269.service) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": "systemd-cryptsetup@luks\\x2db0c50623\\x2de6eb\\x2d411b\\x2db5d3\\x2d25165f6ee269.service", 
    "name": "systemd-cryptsetup@luks\\x2db0c50623\\x2de6eb\\x2d411b\\x2db5d3\\x2d25165f6ee269.service", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "After": "dev-sda.device cryptsetup-pre.target systemd-readahead-replay.service systemd-readahead-collect.service system-systemd\\x2dcryptsetup.slice systemd-journald.socket", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "Before": "cryptsetup.target umount.target", 
        "BindsTo": "dev-sda.device", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "Conflicts": "umount.target", 
        "ControlPID": "0", 
        "DefaultDependencies": "no", 
        "Delegate": "no", 
        "Description": "Cryptography Setup for luks-b0c50623-e6eb-411b-b5d3-25165f6ee269", 
        "DevicePolicy": "auto", 
        "Documentation": "man:crypttab(5) man:systemd-cryptsetup-generator(8) man:systemd-cryptsetup@.service(8)", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-b0c50623-e6eb-411b-b5d3-25165f6ee269 /dev/sda -  ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-b0c50623-e6eb-411b-b5d3-25165f6ee269 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2db0c50623\\x2de6eb\\x2d411b\\x2db5d3\\x2d25165f6ee269.service", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "systemd-cryptsetup@luks\\x2db0c50623\\x2de6eb\\x2d411b\\x2db5d3\\x2d25165f6ee269.service", 
        "IgnoreOnIsolate": "yes", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "4096", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "loaded", 
        "MainPID": "0", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "systemd-cryptsetup@luks\\x2db0c50623\\x2de6eb\\x2d411b\\x2db5d3\\x2d25165f6ee269.service", 
        "NeedDaemonReload": "yes", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "yes", 
        "RequiredBy": "cryptsetup.target", 
        "Requires": "system-systemd\\x2dcryptsetup.slice", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "Slice": "system-systemd\\x2dcryptsetup.slice", 
        "SourcePath": "/etc/crypttab", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "journal", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "0", 
        "TimeoutStopUSec": "0", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "Type": "oneshot", 
        "UMask": "0022", 
        "UnitFilePreset": "disabled", 
        "UnitFileState": "bad", 
        "WantedBy": "dev-sda.device", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  17:56:07 -0500 (0:00:00.508)       0:01:54.982 ***** 
fatal: [managed-node2]: FAILED! => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [], 
    "pools": [], 
    "volumes": []
}

MSG:

cannot remove existing formatting on device 'sda' in safe mode due to adding encryption

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Saturday 14 December 2024  17:56:11 -0500 (0:00:03.663)       0:01:58.646 ***** 
fatal: [managed-node2]: FAILED! => {
    "changed": false
}

MSG:

{u'_ansible_no_log': False, u'crypts': [], u'pools': [], u'leaves': [], u'changed': False, u'actions': [], u'failed': True, u'volumes': [], u'invocation': {u'module_args': {u'packages_only': False, u'disklabel_type': None, u'diskvolume_mkfs_option_map': {u'ext4': u'-F', u'ext3': u'-F', u'ext2': u'-F'}, u'safe_mode': True, u'pools': [], u'volumes': [{u'raid_metadata_version': None, u'raid_level': None, u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': 10737418240, u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'encryption': True, u'mount_device_identifier': u'uuid', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'fs_overwrite_existing': True, u'encryption_key_size': None, u'deduplication': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'cache_size': 0, u'mount_user': None, u'raid_spare_count': None, u'cache_mode': None, u'name': u'foo', u'mount_group': None, u'type': u'disk', u'disks': [u'sda'], u'cached': False, u'thin_pool_size': None, u'thin': None, u'mount_check': 0, u'mount_passno': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}], u'pool_defaults': {u'raid_metadata_version': None, u'encryption_cipher': None, u'encryption_key': None, u'encryption_luks_version': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_password': None, u'encryption': False, u'disks': [], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [], u'shared': False, u'type': u'lvm', u'encryption_key_size': None, u'raid_chunk_size': None}, u'volume_defaults': {u'raid_metadata_version': None, u'raid_level': None, u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': 0, u'mount_point': u'', u'compression': None, u'encryption_password': None, u'encryption': False, u'mount_device_identifier': u'uuid', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'thin_pool_name': None, u'type': u'lvm', u'encryption_key_size': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'cache_size': 0, u'raid_spare_count': None, u'cache_mode': None, u'deduplication': None, u'cached': False, u'fs_overwrite_existing': True, u'disks': [], u'thin_pool_size': None, u'thin': None, u'mount_check': 0, u'mount_passno': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}, u'use_partitions': None}}, u'mounts': [], u'packages': [], u'msg': u"cannot remove existing formatting on device 'sda' in safe mode due to adding encryption"}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  17:56:11 -0500 (0:00:00.069)       0:01:58.716 ***** 
changed: [managed-node2] => (item=systemd-cryptsetup@luks\x2db0c50623\x2de6eb\x2d411b\x2db5d3\x2d25165f6ee269.service) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": "systemd-cryptsetup@luks\\x2db0c50623\\x2de6eb\\x2d411b\\x2db5d3\\x2d25165f6ee269.service", 
    "name": "systemd-cryptsetup@luks\\x2db0c50623\\x2de6eb\\x2d411b\\x2db5d3\\x2d25165f6ee269.service", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "ControlPID": "0", 
        "DefaultDependencies": "yes", 
        "Delegate": "no", 
        "Description": "systemd-cryptsetup@luks\\x2db0c50623\\x2de6eb\\x2d411b\\x2db5d3\\x2d25165f6ee269.service", 
        "DevicePolicy": "auto", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/dev/null", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "systemd-cryptsetup@luks\\x2db0c50623\\x2de6eb\\x2d411b\\x2db5d3\\x2d25165f6ee269.service", 
        "IgnoreOnIsolate": "no", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "1048576", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "masked", 
        "MainPID": "0", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "systemd-cryptsetup@luks\\x2db0c50623\\x2de6eb\\x2d411b\\x2db5d3\\x2d25165f6ee269.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "no", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "inherit", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "1min 30s", 
        "TimeoutStopUSec": "1min 30s", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "UMask": "0022", 
        "UnitFileState": "bad", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Saturday 14 December 2024  17:56:11 -0500 (0:00:00.610)       0:01:59.327 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Saturday 14 December 2024  17:56:12 -0500 (0:00:00.071)       0:01:59.398 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Saturday 14 December 2024  17:56:12 -0500 (0:00:00.086)       0:01:59.484 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Stat the file] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:11
Saturday 14 December 2024  17:56:12 -0500 (0:00:00.059)       0:01:59.544 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734216958.0008383, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", 
        "ctime": 1734216958.0008383, 
        "dev": 2048, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 67, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/x-empty", 
        "mode": "0644", 
        "mtime": 1734216958.0008383, 
        "nlink": 1, 
        "path": "/opt/test1/quux", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": "18446744071879105950", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Assert file presence] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:16
Saturday 14 December 2024  17:56:12 -0500 (0:00:00.471)       0:02:00.015 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Add encryption to the volume] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:177
Saturday 14 December 2024  17:56:12 -0500 (0:00:00.062)       0:02:00.078 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  17:56:12 -0500 (0:00:00.250)       0:02:00.328 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  17:56:13 -0500 (0:00:00.166)       0:02:00.494 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  17:56:13 -0500 (0:00:00.063)       0:02:00.558 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  17:56:13 -0500 (0:00:00.137)       0:02:00.696 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  17:56:13 -0500 (0:00:00.080)       0:02:00.777 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  17:56:13 -0500 (0:00:00.098)       0:02:00.875 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  17:56:13 -0500 (0:00:00.096)       0:02:00.972 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  17:56:13 -0500 (0:00:00.122)       0:02:01.095 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  17:56:13 -0500 (0:00:00.172)       0:02:01.267 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  17:56:15 -0500 (0:00:01.305)       0:02:02.573 ***** 
ok: [managed-node2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  17:56:15 -0500 (0:00:00.061)       0:02:02.635 ***** 
ok: [managed-node2] => {
    "storage_volumes": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": true, 
            "encryption_password": "yabbadabbadoo", 
            "mount_point": "/opt/test1", 
            "name": "foo", 
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  17:56:15 -0500 (0:00:00.068)       0:02:02.703 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [
        "cryptsetup"
    ], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  17:56:19 -0500 (0:00:03.927)       0:02:06.631 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  17:56:19 -0500 (0:00:00.103)       0:02:06.734 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  17:56:19 -0500 (0:00:00.107)       0:02:06.841 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  17:56:19 -0500 (0:00:00.061)       0:02:06.902 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  17:56:19 -0500 (0:00:00.065)       0:02:06.968 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed", 
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: cryptsetup kpartx

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  17:56:20 -0500 (0:00:00.877)       0:02:07.846 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  17:56:21 -0500 (0:00:01.119)       0:02:08.965 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  17:56:21 -0500 (0:00:00.097)       0:02:09.063 ***** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  17:56:21 -0500 (0:00:00.055)       0:02:09.118 ***** 
changed: [managed-node2] => {
    "actions": [
        {
            "action": "destroy format", 
            "device": "/dev/sda", 
            "fs_type": "xfs"
        }, 
        {
            "action": "create format", 
            "device": "/dev/sda", 
            "fs_type": "luks"
        }, 
        {
            "action": "create device", 
            "device": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
            "fs_type": null
        }, 
        {
            "action": "create format", 
            "device": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
            "fs_type": "xfs"
        }
    ], 
    "changed": true, 
    "crypts": [
        {
            "backing_device": "/dev/sda", 
            "name": "luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
            "password": "-", 
            "state": "present"
        }
    ], 
    "leaves": [
        "/dev/sdb", 
        "/dev/sdc", 
        "/dev/sdd", 
        "/dev/sde", 
        "/dev/sdf", 
        "/dev/sdg", 
        "/dev/sdh", 
        "/dev/sdi", 
        "/dev/xvda1", 
        "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667"
    ], 
    "mounts": [
        {
            "fstype": "xfs", 
            "path": "/opt/test1", 
            "src": "UUID=6b177eed-d54c-42ab-b8d1-951cc9acb8aa", 
            "state": "absent"
        }, 
        {
            "dump": 0, 
            "fstype": "xfs", 
            "group": null, 
            "mode": null, 
            "opts": "defaults", 
            "owner": null, 
            "passno": 0, 
            "path": "/opt/test1", 
            "src": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
            "state": "mounted"
        }
    ], 
    "packages": [
        "cryptsetup", 
        "xfsprogs", 
        "e2fsprogs"
    ], 
    "pools": [], 
    "volumes": [
        {
            "_device": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
            "_kernel_device": "/dev/dm-0", 
            "_mount_id": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
            "_raw_device": "/dev/sda", 
            "_raw_kernel_device": "/dev/sda", 
            "cache_devices": [], 
            "cache_mode": null, 
            "cache_size": 0, 
            "cached": false, 
            "compression": null, 
            "deduplication": null, 
            "disks": [
                "sda"
            ], 
            "encryption": true, 
            "encryption_cipher": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
            "fs_create_options": "", 
            "fs_label": "", 
            "fs_overwrite_existing": true, 
            "fs_type": "xfs", 
            "mount_check": 0, 
            "mount_device_identifier": "uuid", 
            "mount_group": null, 
            "mount_mode": null, 
            "mount_options": "defaults", 
            "mount_passno": 0, 
            "mount_point": "/opt/test1", 
            "mount_user": null, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "raid_stripe_size": null, 
            "size": 10737418240, 
            "state": "present", 
            "thin": null, 
            "thin_pool_name": null, 
            "thin_pool_size": null, 
            "type": "disk", 
            "vdo_pool_size": null
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Saturday 14 December 2024  17:56:32 -0500 (0:00:10.357)       0:02:19.475 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Saturday 14 December 2024  17:56:32 -0500 (0:00:00.053)       0:02:19.529 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734216944.5538445, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "668cc2654ae463ae9d1cc60621c2706ebf42dcdd", 
        "ctime": 1734216944.5508444, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 263588, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0644", 
        "mtime": 1734216944.5508444, 
        "nlink": 1, 
        "path": "/etc/fstab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 1299, 
        "uid": 0, 
        "version": "606402765", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Saturday 14 December 2024  17:56:32 -0500 (0:00:00.473)       0:02:20.002 ***** 
ok: [managed-node2] => {
    "backup": "", 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  17:56:33 -0500 (0:00:00.452)       0:02:20.455 ***** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Saturday 14 December 2024  17:56:33 -0500 (0:00:00.053)       0:02:20.508 ***** 
ok: [managed-node2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format", 
                "device": "/dev/sda", 
                "fs_type": "xfs"
            }, 
            {
                "action": "create format", 
                "device": "/dev/sda", 
                "fs_type": "luks"
            }, 
            {
                "action": "create device", 
                "device": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
                "fs_type": null
            }, 
            {
                "action": "create format", 
                "device": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
                "fs_type": "xfs"
            }
        ], 
        "changed": true, 
        "crypts": [
            {
                "backing_device": "/dev/sda", 
                "name": "luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
                "password": "-", 
                "state": "present"
            }
        ], 
        "failed": false, 
        "leaves": [
            "/dev/sdb", 
            "/dev/sdc", 
            "/dev/sdd", 
            "/dev/sde", 
            "/dev/sdf", 
            "/dev/sdg", 
            "/dev/sdh", 
            "/dev/sdi", 
            "/dev/xvda1", 
            "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667"
        ], 
        "mounts": [
            {
                "fstype": "xfs", 
                "path": "/opt/test1", 
                "src": "UUID=6b177eed-d54c-42ab-b8d1-951cc9acb8aa", 
                "state": "absent"
            }, 
            {
                "dump": 0, 
                "fstype": "xfs", 
                "group": null, 
                "mode": null, 
                "opts": "defaults", 
                "owner": null, 
                "passno": 0, 
                "path": "/opt/test1", 
                "src": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
                "state": "mounted"
            }
        ], 
        "packages": [
            "cryptsetup", 
            "xfsprogs", 
            "e2fsprogs"
        ], 
        "pools": [], 
        "volumes": [
            {
                "_device": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
                "_kernel_device": "/dev/dm-0", 
                "_mount_id": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
                "_raw_device": "/dev/sda", 
                "_raw_kernel_device": "/dev/sda", 
                "cache_devices": [], 
                "cache_mode": null, 
                "cache_size": 0, 
                "cached": false, 
                "compression": null, 
                "deduplication": null, 
                "disks": [
                    "sda"
                ], 
                "encryption": true, 
                "encryption_cipher": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                "fs_create_options": "", 
                "fs_label": "", 
                "fs_overwrite_existing": true, 
                "fs_type": "xfs", 
                "mount_check": 0, 
                "mount_device_identifier": "uuid", 
                "mount_group": null, 
                "mount_mode": null, 
                "mount_options": "defaults", 
                "mount_passno": 0, 
                "mount_point": "/opt/test1", 
                "mount_user": null, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "raid_stripe_size": null, 
                "size": 10737418240, 
                "state": "present", 
                "thin": null, 
                "thin_pool_name": null, 
                "thin_pool_size": null, 
                "type": "disk", 
                "vdo_pool_size": null
            }
        ]
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Saturday 14 December 2024  17:56:33 -0500 (0:00:00.094)       0:02:20.603 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Saturday 14 December 2024  17:56:33 -0500 (0:00:00.060)       0:02:20.663 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": [
            {
                "_device": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
                "_kernel_device": "/dev/dm-0", 
                "_mount_id": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
                "_raw_device": "/dev/sda", 
                "_raw_kernel_device": "/dev/sda", 
                "cache_devices": [], 
                "cache_mode": null, 
                "cache_size": 0, 
                "cached": false, 
                "compression": null, 
                "deduplication": null, 
                "disks": [
                    "sda"
                ], 
                "encryption": true, 
                "encryption_cipher": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                "fs_create_options": "", 
                "fs_label": "", 
                "fs_overwrite_existing": true, 
                "fs_type": "xfs", 
                "mount_check": 0, 
                "mount_device_identifier": "uuid", 
                "mount_group": null, 
                "mount_mode": null, 
                "mount_options": "defaults", 
                "mount_passno": 0, 
                "mount_point": "/opt/test1", 
                "mount_user": null, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "raid_stripe_size": null, 
                "size": 10737418240, 
                "state": "present", 
                "thin": null, 
                "thin_pool_name": null, 
                "thin_pool_size": null, 
                "type": "disk", 
                "vdo_pool_size": null
            }
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Saturday 14 December 2024  17:56:33 -0500 (0:00:00.064)       0:02:20.727 ***** 
changed: [managed-node2] => (item={u'src': u'UUID=6b177eed-d54c-42ab-b8d1-951cc9acb8aa', u'state': u'absent', u'fstype': u'xfs', u'path': u'/opt/test1'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": true, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "fstype": "xfs", 
        "path": "/opt/test1", 
        "src": "UUID=6b177eed-d54c-42ab-b8d1-951cc9acb8aa", 
        "state": "absent"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "UUID=6b177eed-d54c-42ab-b8d1-951cc9acb8aa"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Saturday 14 December 2024  17:56:33 -0500 (0:00:00.582)       0:02:21.310 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Saturday 14 December 2024  17:56:34 -0500 (0:00:00.640)       0:02:21.951 ***** 
changed: [managed-node2] => (item={u'src': u'/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": true, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
        "state": "mounted"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Saturday 14 December 2024  17:56:35 -0500 (0:00:00.452)       0:02:22.403 ***** 
skipping: [managed-node2] => (item={u'src': u'/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'})  => {
    "ansible_loop_var": "mount_info", 
    "changed": false, 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
        "state": "mounted"
    }, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Saturday 14 December 2024  17:56:35 -0500 (0:00:00.068)       0:02:22.472 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Saturday 14 December 2024  17:56:35 -0500 (0:00:00.672)       0:02:23.144 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734216948.9888425, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", 
        "ctime": 1734216946.2088437, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 917510, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/x-empty", 
        "mode": "0600", 
        "mtime": 1734216946.2088437, 
        "nlink": 1, 
        "path": "/etc/crypttab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": false, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": "606408720", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Saturday 14 December 2024  17:56:36 -0500 (0:00:00.414)       0:02:23.559 ***** 
changed: [managed-node2] => (item={u'state': u'present', u'password': u'-', u'name': u'luks-b832bb64-70e9-412c-8abe-83daccf8b667', u'backing_device': u'/dev/sda'}) => {
    "ansible_loop_var": "entry", 
    "backup": "", 
    "changed": true, 
    "entry": {
        "backing_device": "/dev/sda", 
        "name": "luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
        "password": "-", 
        "state": "present"
    }
}

MSG:

line added

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Saturday 14 December 2024  17:56:36 -0500 (0:00:00.470)       0:02:24.029 ***** 
ok: [managed-node2]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:190
Saturday 14 December 2024  17:56:37 -0500 (0:00:00.987)       0:02:25.017 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node2

TASK [Print out pool information] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Saturday 14 December 2024  17:56:37 -0500 (0:00:00.301)       0:02:25.318 ***** 
skipping: [managed-node2] => {}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Saturday 14 December 2024  17:56:38 -0500 (0:00:00.095)       0:02:25.413 ***** 
ok: [managed-node2] => {
    "_storage_volumes_list": [
        {
            "_device": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
            "_kernel_device": "/dev/dm-0", 
            "_mount_id": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
            "_raw_device": "/dev/sda", 
            "_raw_kernel_device": "/dev/sda", 
            "cache_devices": [], 
            "cache_mode": null, 
            "cache_size": 0, 
            "cached": false, 
            "compression": null, 
            "deduplication": null, 
            "disks": [
                "sda"
            ], 
            "encryption": true, 
            "encryption_cipher": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
            "fs_create_options": "", 
            "fs_label": "", 
            "fs_overwrite_existing": true, 
            "fs_type": "xfs", 
            "mount_check": 0, 
            "mount_device_identifier": "uuid", 
            "mount_group": null, 
            "mount_mode": null, 
            "mount_options": "defaults", 
            "mount_passno": 0, 
            "mount_point": "/opt/test1", 
            "mount_user": null, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "raid_stripe_size": null, 
            "size": 10737418240, 
            "state": "present", 
            "thin": null, 
            "thin_pool_name": null, 
            "thin_pool_size": null, 
            "type": "disk", 
            "vdo_pool_size": null
        }
    ]
}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Saturday 14 December 2024  17:56:38 -0500 (0:00:00.096)       0:02:25.510 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "info": {
        "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667": {
            "fstype": "xfs", 
            "label": "", 
            "mountpoint": "/opt/test1", 
            "name": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
            "size": "10G", 
            "type": "crypt", 
            "uuid": "c3f3cf38-a864-42f8-b514-115f90268b09"
        }, 
        "/dev/sda": {
            "fstype": "crypto_LUKS", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sda", 
            "size": "10G", 
            "type": "disk", 
            "uuid": "b832bb64-70e9-412c-8abe-83daccf8b667"
        }, 
        "/dev/sdb": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdb", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdc": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdc", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdd": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdd", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sde": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sde", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdf": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdf", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdg": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdg", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdh": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdh", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdi": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdi", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/xvda", 
            "size": "250G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda1": {
            "fstype": "ext4", 
            "label": "", 
            "mountpoint": "/", 
            "name": "/dev/xvda1", 
            "size": "250G", 
            "type": "partition", 
            "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Saturday 14 December 2024  17:56:38 -0500 (0:00:00.572)       0:02:26.082 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/fstab"
    ], 
    "delta": "0:00:00.003386", 
    "end": "2024-12-14 17:56:39.088676", 
    "rc": 0, 
    "start": "2024-12-14 17:56:39.085290"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Thu Jun 20 10:23:46 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 /                       ext4    defaults        1 1
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667 /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Saturday 14 December 2024  17:56:39 -0500 (0:00:00.471)       0:02:26.553 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/crypttab"
    ], 
    "delta": "0:00:00.003387", 
    "end": "2024-12-14 17:56:39.582306", 
    "failed_when_result": false, 
    "rc": 0, 
    "start": "2024-12-14 17:56:39.578919"
}

STDOUT:

luks-b832bb64-70e9-412c-8abe-83daccf8b667 /dev/sda -

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Saturday 14 December 2024  17:56:39 -0500 (0:00:00.488)       0:02:27.042 ***** 

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Saturday 14 December 2024  17:56:39 -0500 (0:00:00.049)       0:02:27.091 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node2

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Saturday 14 December 2024  17:56:39 -0500 (0:00:00.124)       0:02:27.216 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true, 
        "_storage_volume_tests": [
            "mount", 
            "fstab", 
            "fs", 
            "device", 
            "encryption", 
            "md", 
            "size", 
            "cache"
        ]
    }, 
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Saturday 14 December 2024  17:56:39 -0500 (0:00:00.068)       0:02:27.285 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node2

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Saturday 14 December 2024  17:56:40 -0500 (0:00:00.429)       0:02:27.715 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667"
    }, 
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Saturday 14 December 2024  17:56:40 -0500 (0:00:00.109)       0:02:27.824 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1", 
        "storage_test_swap_expected_matches": "0"
    }, 
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Saturday 14 December 2024  17:56:40 -0500 (0:00:00.095)       0:02:27.920 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Saturday 14 December 2024  17:56:40 -0500 (0:00:00.055)       0:02:27.975 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Saturday 14 December 2024  17:56:40 -0500 (0:00:00.095)       0:02:28.071 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Saturday 14 December 2024  17:56:40 -0500 (0:00:00.082)       0:02:28.154 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Saturday 14 December 2024  17:56:40 -0500 (0:00:00.084)       0:02:28.238 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Saturday 14 December 2024  17:56:40 -0500 (0:00:00.091)       0:02:28.330 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Saturday 14 December 2024  17:56:41 -0500 (0:00:00.087)       0:02:28.417 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Saturday 14 December 2024  17:56:41 -0500 (0:00:00.057)       0:02:28.474 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Saturday 14 December 2024  17:56:41 -0500 (0:00:00.128)       0:02:28.603 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null, 
        "storage_test_mount_expected_mount_point": null, 
        "storage_test_swap_expected_matches": null, 
        "storage_test_swaps": null, 
        "storage_test_sys_node": null
    }, 
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Saturday 14 December 2024  17:56:41 -0500 (0:00:00.119)       0:02:28.723 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1", 
        "storage_test_fstab_expected_mount_options_matches": "1", 
        "storage_test_fstab_expected_mount_point_matches": "1", 
        "storage_test_fstab_id_matches": [
            "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667 "
        ], 
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ], 
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    }, 
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Saturday 14 December 2024  17:56:41 -0500 (0:00:00.090)       0:02:28.814 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Saturday 14 December 2024  17:56:41 -0500 (0:00:00.068)       0:02:28.882 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Saturday 14 December 2024  17:56:41 -0500 (0:00:00.060)       0:02:28.943 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Saturday 14 December 2024  17:56:41 -0500 (0:00:00.053)       0:02:28.997 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Saturday 14 December 2024  17:56:41 -0500 (0:00:00.064)       0:02:29.062 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null, 
        "storage_test_fstab_expected_mount_options_matches": null, 
        "storage_test_fstab_expected_mount_point_matches": null, 
        "storage_test_fstab_id_matches": null, 
        "storage_test_fstab_mount_options_matches": null, 
        "storage_test_fstab_mount_point_matches": null
    }, 
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Saturday 14 December 2024  17:56:41 -0500 (0:00:00.151)       0:02:29.213 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Saturday 14 December 2024  17:56:41 -0500 (0:00:00.143)       0:02:29.357 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Saturday 14 December 2024  17:56:42 -0500 (0:00:00.117)       0:02:29.475 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734216991.8298225, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "ctime": 1734216991.8298225, 
        "dev": 5, 
        "device_type": 2048, 
        "executable": false, 
        "exists": true, 
        "gid": 6, 
        "gr_name": "disk", 
        "inode": 28408, 
        "isblk": true, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": false, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/blockdevice", 
        "mode": "0660", 
        "mtime": 1734216991.8298225, 
        "nlink": 1, 
        "path": "/dev/sda", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": null, 
        "wgrp": true, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Saturday 14 December 2024  17:56:42 -0500 (0:00:00.501)       0:02:29.976 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Saturday 14 December 2024  17:56:42 -0500 (0:00:00.146)       0:02:30.123 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Saturday 14 December 2024  17:56:42 -0500 (0:00:00.100)       0:02:30.224 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Saturday 14 December 2024  17:56:42 -0500 (0:00:00.117)       0:02:30.341 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "st_volume_type": "disk"
    }, 
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Saturday 14 December 2024  17:56:43 -0500 (0:00:00.129)       0:02:30.470 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Saturday 14 December 2024  17:56:43 -0500 (0:00:00.106)       0:02:30.577 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Saturday 14 December 2024  17:56:43 -0500 (0:00:00.142)       0:02:30.719 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734216991.9408226, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "ctime": 1734216991.9408226, 
        "dev": 5, 
        "device_type": 64768, 
        "executable": false, 
        "exists": true, 
        "gid": 6, 
        "gr_name": "disk", 
        "inode": 391751, 
        "isblk": true, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": false, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/symlink", 
        "mode": "0660", 
        "mtime": 1734216991.9408226, 
        "nlink": 1, 
        "path": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": null, 
        "wgrp": true, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Saturday 14 December 2024  17:56:44 -0500 (0:00:00.712)       0:02:31.432 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}
lsrpackages: cryptsetup

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Saturday 14 December 2024  17:56:44 -0500 (0:00:00.786)       0:02:32.218 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cryptsetup", 
        "luksDump", 
        "/dev/sda"
    ], 
    "delta": "0:00:00.026665", 
    "end": "2024-12-14 17:56:45.347474", 
    "rc": 0, 
    "start": "2024-12-14 17:56:45.320809"
}

STDOUT:

LUKS header information for /dev/sda

Version:       	1
Cipher name:   	aes
Cipher mode:   	xts-plain64
Hash spec:     	sha256
Payload offset:	8192
MK bits:       	512
MK digest:     	07 f0 44 6a d9 fd c7 ea a0 92 16 c9 fd 35 39 78 56 6f 11 5d 
MK salt:       	24 f5 6d e5 23 87 34 e0 7c b7 f2 39 24 02 00 a8 
               	67 1f d9 86 3b 95 a5 d8 a7 72 1e 67 95 8c dd 8c 
MK iterations: 	23076
UUID:          	b832bb64-70e9-412c-8abe-83daccf8b667

Key Slot 0: ENABLED
	Iterations:         	368178
	Salt:               	d9 07 24 fb 03 08 26 36 fe ff ec 07 32 e4 1f 31 
	                      	3c cd 00 06 e7 32 a1 7e f6 45 76 6a bf 95 b6 78 
	Key material offset:	8
	AF stripes:            	4000
Key Slot 1: DISABLED
Key Slot 2: DISABLED
Key Slot 3: DISABLED
Key Slot 4: DISABLED
Key Slot 5: DISABLED
Key Slot 6: DISABLED
Key Slot 7: DISABLED

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Saturday 14 December 2024  17:56:45 -0500 (0:00:00.572)       0:02:32.791 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Saturday 14 December 2024  17:56:45 -0500 (0:00:00.046)       0:02:32.837 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Saturday 14 December 2024  17:56:45 -0500 (0:00:00.052)       0:02:32.890 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Saturday 14 December 2024  17:56:45 -0500 (0:00:00.058)       0:02:32.949 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Saturday 14 December 2024  17:56:45 -0500 (0:00:00.103)       0:02:33.052 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Saturday 14 December 2024  17:56:45 -0500 (0:00:00.079)       0:02:33.131 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Saturday 14 December 2024  17:56:45 -0500 (0:00:00.094)       0:02:33.226 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Saturday 14 December 2024  17:56:45 -0500 (0:00:00.077)       0:02:33.303 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [
            "luks-b832bb64-70e9-412c-8abe-83daccf8b667 /dev/sda -"
        ], 
        "_storage_test_expected_crypttab_entries": "1", 
        "_storage_test_expected_crypttab_key_file": "-"
    }, 
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Saturday 14 December 2024  17:56:46 -0500 (0:00:00.117)       0:02:33.420 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Saturday 14 December 2024  17:56:46 -0500 (0:00:00.098)       0:02:33.518 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Saturday 14 December 2024  17:56:46 -0500 (0:00:00.103)       0:02:33.622 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Saturday 14 December 2024  17:56:46 -0500 (0:00:00.133)       0:02:33.756 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Saturday 14 December 2024  17:56:46 -0500 (0:00:00.100)       0:02:33.856 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null, 
        "_storage_test_expected_crypttab_entries": null, 
        "_storage_test_expected_crypttab_key_file": null
    }, 
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Saturday 14 December 2024  17:56:46 -0500 (0:00:00.060)       0:02:33.917 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Saturday 14 December 2024  17:56:46 -0500 (0:00:00.061)       0:02:33.979 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Saturday 14 December 2024  17:56:46 -0500 (0:00:00.056)       0:02:34.035 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Saturday 14 December 2024  17:56:46 -0500 (0:00:00.074)       0:02:34.110 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Saturday 14 December 2024  17:56:46 -0500 (0:00:00.062)       0:02:34.173 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Saturday 14 December 2024  17:56:46 -0500 (0:00:00.067)       0:02:34.240 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Saturday 14 December 2024  17:56:46 -0500 (0:00:00.053)       0:02:34.293 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Saturday 14 December 2024  17:56:46 -0500 (0:00:00.052)       0:02:34.345 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Saturday 14 December 2024  17:56:47 -0500 (0:00:00.052)       0:02:34.398 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Saturday 14 December 2024  17:56:47 -0500 (0:00:00.055)       0:02:34.454 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Saturday 14 December 2024  17:56:47 -0500 (0:00:00.052)       0:02:34.506 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Saturday 14 December 2024  17:56:47 -0500 (0:00:00.074)       0:02:34.581 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Saturday 14 December 2024  17:56:47 -0500 (0:00:00.061)       0:02:34.642 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Saturday 14 December 2024  17:56:47 -0500 (0:00:00.057)       0:02:34.700 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Saturday 14 December 2024  17:56:47 -0500 (0:00:00.058)       0:02:34.758 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Saturday 14 December 2024  17:56:47 -0500 (0:00:00.078)       0:02:34.837 ***** 
skipping: [managed-node2] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Saturday 14 December 2024  17:56:47 -0500 (0:00:00.055)       0:02:34.892 ***** 
skipping: [managed-node2] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Saturday 14 December 2024  17:56:47 -0500 (0:00:00.097)       0:02:34.990 ***** 
skipping: [managed-node2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Saturday 14 December 2024  17:56:47 -0500 (0:00:00.072)       0:02:35.062 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Saturday 14 December 2024  17:56:47 -0500 (0:00:00.080)       0:02:35.142 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Saturday 14 December 2024  17:56:47 -0500 (0:00:00.069)       0:02:35.212 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Saturday 14 December 2024  17:56:47 -0500 (0:00:00.061)       0:02:35.274 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Saturday 14 December 2024  17:56:47 -0500 (0:00:00.068)       0:02:35.343 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Saturday 14 December 2024  17:56:48 -0500 (0:00:00.055)       0:02:35.398 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Saturday 14 December 2024  17:56:48 -0500 (0:00:00.051)       0:02:35.450 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Saturday 14 December 2024  17:56:48 -0500 (0:00:00.062)       0:02:35.512 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Saturday 14 December 2024  17:56:48 -0500 (0:00:00.054)       0:02:35.566 ***** 
skipping: [managed-node2] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Saturday 14 December 2024  17:56:48 -0500 (0:00:00.128)       0:02:35.695 ***** 
skipping: [managed-node2] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Saturday 14 December 2024  17:56:48 -0500 (0:00:00.055)       0:02:35.750 ***** 
skipping: [managed-node2] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Saturday 14 December 2024  17:56:48 -0500 (0:00:00.060)       0:02:35.811 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Saturday 14 December 2024  17:56:48 -0500 (0:00:00.069)       0:02:35.881 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Saturday 14 December 2024  17:56:48 -0500 (0:00:00.053)       0:02:35.935 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Saturday 14 December 2024  17:56:48 -0500 (0:00:00.059)       0:02:35.995 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Saturday 14 December 2024  17:56:48 -0500 (0:00:00.059)       0:02:36.054 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Saturday 14 December 2024  17:56:48 -0500 (0:00:00.051)       0:02:36.106 ***** 
ok: [managed-node2] => {
    "storage_test_actual_size": {
        "changed": false, 
        "skip_reason": "Conditional result was False", 
        "skipped": true
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Saturday 14 December 2024  17:56:48 -0500 (0:00:00.057)       0:02:36.164 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Saturday 14 December 2024  17:56:48 -0500 (0:00:00.083)       0:02:36.247 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Saturday 14 December 2024  17:56:48 -0500 (0:00:00.071)       0:02:36.319 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Saturday 14 December 2024  17:56:49 -0500 (0:00:00.069)       0:02:36.388 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Saturday 14 December 2024  17:56:49 -0500 (0:00:00.065)       0:02:36.454 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Saturday 14 December 2024  17:56:49 -0500 (0:00:00.053)       0:02:36.508 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Saturday 14 December 2024  17:56:49 -0500 (0:00:00.054)       0:02:36.562 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Saturday 14 December 2024  17:56:49 -0500 (0:00:00.055)       0:02:36.618 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Saturday 14 December 2024  17:56:49 -0500 (0:00:00.051)       0:02:36.669 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Saturday 14 December 2024  17:56:49 -0500 (0:00:00.054)       0:02:36.724 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    }, 
    "changed": false
}

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Saturday 14 December 2024  17:56:49 -0500 (0:00:00.057)       0:02:36.781 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null, 
        "storage_test_crypttab": null, 
        "storage_test_fstab": null
    }, 
    "changed": false
}

TASK [Test for correct handling of new encrypted volume w/ no key] *************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:197
Saturday 14 December 2024  17:56:49 -0500 (0:00:00.054)       0:02:36.836 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node2

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Saturday 14 December 2024  17:56:49 -0500 (0:00:00.117)       0:02:36.953 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_pools_global": [], 
        "storage_safe_mode_global": true, 
        "storage_volumes_global": []
    }, 
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Saturday 14 December 2024  17:56:49 -0500 (0:00:00.138)       0:02:37.092 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  17:56:49 -0500 (0:00:00.118)       0:02:37.211 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  17:56:49 -0500 (0:00:00.126)       0:02:37.338 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  17:56:50 -0500 (0:00:00.072)       0:02:37.411 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  17:56:50 -0500 (0:00:00.144)       0:02:37.556 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  17:56:50 -0500 (0:00:00.053)       0:02:37.610 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  17:56:50 -0500 (0:00:00.104)       0:02:37.714 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  17:56:50 -0500 (0:00:00.143)       0:02:37.858 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  17:56:50 -0500 (0:00:00.092)       0:02:37.951 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  17:56:50 -0500 (0:00:00.155)       0:02:38.107 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  17:56:52 -0500 (0:00:01.529)       0:02:39.636 ***** 
ok: [managed-node2] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ], 
            "name": "foo", 
            "type": "partition", 
            "volumes": [
                {
                    "encryption": true, 
                    "mount_point": "/opt/test1", 
                    "name": "test1", 
                    "size": "4g", 
                    "type": "partition"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  17:56:52 -0500 (0:00:00.098)       0:02:39.735 ***** 
ok: [managed-node2] => {
    "storage_volumes": []
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  17:56:52 -0500 (0:00:00.096)       0:02:39.832 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [
        "cryptsetup"
    ], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  17:56:56 -0500 (0:00:03.753)       0:02:43.585 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  17:56:56 -0500 (0:00:00.074)       0:02:43.660 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  17:56:56 -0500 (0:00:00.033)       0:02:43.693 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  17:56:56 -0500 (0:00:00.037)       0:02:43.731 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  17:56:56 -0500 (0:00:00.033)       0:02:43.764 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed", 
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: cryptsetup kpartx

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  17:56:57 -0500 (0:00:00.624)       0:02:44.389 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  17:56:57 -0500 (0:00:00.952)       0:02:45.341 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  17:56:58 -0500 (0:00:00.055)       0:02:45.397 ***** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  17:56:58 -0500 (0:00:00.033)       0:02:45.430 ***** 
fatal: [managed-node2]: FAILED! => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [], 
    "pools": [], 
    "volumes": []
}

MSG:

encrypted volume 'test1' missing key/password

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Saturday 14 December 2024  17:57:02 -0500 (0:00:04.073)       0:02:49.503 ***** 
fatal: [managed-node2]: FAILED! => {
    "changed": false
}

MSG:

{u'_ansible_no_log': False, u'crypts': [], u'pools': [], u'leaves': [], u'changed': False, u'actions': [], u'failed': True, u'volumes': [], u'invocation': {u'module_args': {u'packages_only': False, u'disklabel_type': None, u'diskvolume_mkfs_option_map': {u'ext4': u'-F', u'ext3': u'-F', u'ext2': u'-F'}, u'safe_mode': False, u'pools': [{u'raid_metadata_version': None, u'encryption_key_size': None, u'encryption_key': None, u'encryption_luks_version': None, u'encryption_tang_url': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_tang_thumbprint': None, u'name': u'foo', u'encryption_password': None, u'encryption': False, u'disks': [u'sda'], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [{u'raid_metadata_version': None, u'mount_device_identifier': u'uuid', u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': u'4g', u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': None, u'encryption': True, u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'type': u'partition', u'encryption_key_size': None, u'deduplication': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'mount_passno': 0, u'mount_user': None, u'raid_spare_count': None, u'name': u'test1', u'cache_mode': None, u'raid_disks': [], u'mount_group': None, u'fs_overwrite_existing': True, u'disks': [], u'cached': False, u'thin_pool_size': None, u'thin': False, u'mount_check': 0, u'cache_size': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}], u'shared': False, u'encryption_clevis_pin': None, u'type': u'partition', u'encryption_cipher': None, u'raid_chunk_size': None}], u'volumes': [], u'pool_defaults': {u'raid_metadata_version': None, u'encryption_cipher': None, u'encryption_key': None, u'encryption_luks_version': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_password': None, u'encryption': False, u'disks': [], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [], u'shared': False, u'type': u'lvm', u'encryption_key_size': None, u'raid_chunk_size': None}, u'volume_defaults': {u'raid_metadata_version': None, u'raid_level': None, u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': 0, u'mount_point': u'', u'compression': None, u'encryption_password': None, u'encryption': False, u'mount_device_identifier': u'uuid', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'thin_pool_name': None, u'type': u'lvm', u'encryption_key_size': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'cache_size': 0, u'raid_spare_count': None, u'cache_mode': None, u'deduplication': None, u'cached': False, u'fs_overwrite_existing': True, u'disks': [], u'thin_pool_size': None, u'thin': None, u'mount_check': 0, u'mount_passno': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}, u'use_partitions': None}}, u'mounts': [], u'packages': [], u'msg': u"encrypted volume 'test1' missing key/password"}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  17:57:02 -0500 (0:00:00.094)       0:02:49.598 ***** 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Saturday 14 December 2024  17:57:02 -0500 (0:00:00.050)       0:02:49.649 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Saturday 14 December 2024  17:57:02 -0500 (0:00:00.069)       0:02:49.718 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Saturday 14 December 2024  17:57:02 -0500 (0:00:00.072)       0:02:49.791 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Create an encrypted partition volume w/ default fs] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:216
Saturday 14 December 2024  17:57:02 -0500 (0:00:00.065)       0:02:49.856 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  17:57:02 -0500 (0:00:00.232)       0:02:50.089 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  17:57:02 -0500 (0:00:00.062)       0:02:50.152 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  17:57:02 -0500 (0:00:00.059)       0:02:50.211 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  17:57:02 -0500 (0:00:00.130)       0:02:50.341 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  17:57:03 -0500 (0:00:00.052)       0:02:50.394 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  17:57:03 -0500 (0:00:00.053)       0:02:50.448 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  17:57:03 -0500 (0:00:00.057)       0:02:50.506 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  17:57:03 -0500 (0:00:00.055)       0:02:50.561 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  17:57:03 -0500 (0:00:00.129)       0:02:50.691 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  17:57:04 -0500 (0:00:01.408)       0:02:52.099 ***** 
ok: [managed-node2] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ], 
            "name": "foo", 
            "type": "partition", 
            "volumes": [
                {
                    "encryption": true, 
                    "encryption_password": "yabbadabbadoo", 
                    "mount_point": "/opt/test1", 
                    "name": "test1", 
                    "size": "4g", 
                    "type": "partition"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  17:57:04 -0500 (0:00:00.064)       0:02:52.164 ***** 
ok: [managed-node2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  17:57:04 -0500 (0:00:00.065)       0:02:52.229 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [
        "cryptsetup"
    ], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  17:57:09 -0500 (0:00:04.206)       0:02:56.436 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  17:57:09 -0500 (0:00:00.122)       0:02:56.558 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  17:57:09 -0500 (0:00:00.065)       0:02:56.624 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  17:57:09 -0500 (0:00:00.053)       0:02:56.677 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  17:57:09 -0500 (0:00:00.052)       0:02:56.729 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed", 
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: cryptsetup kpartx

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  17:57:10 -0500 (0:00:01.033)       0:02:57.763 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  17:57:11 -0500 (0:00:01.160)       0:02:58.923 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  17:57:11 -0500 (0:00:00.111)       0:02:59.035 ***** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  17:57:11 -0500 (0:00:00.088)       0:02:59.123 ***** 
changed: [managed-node2] => {
    "actions": [
        {
            "action": "destroy format", 
            "device": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
            "fs_type": "xfs"
        }, 
        {
            "action": "destroy device", 
            "device": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
            "fs_type": null
        }, 
        {
            "action": "destroy format", 
            "device": "/dev/sda", 
            "fs_type": "luks"
        }, 
        {
            "action": "create format", 
            "device": "/dev/sda", 
            "fs_type": "disklabel"
        }, 
        {
            "action": "create device", 
            "device": "/dev/sda1", 
            "fs_type": null
        }, 
        {
            "action": "create format", 
            "device": "/dev/sda1", 
            "fs_type": "luks"
        }, 
        {
            "action": "create device", 
            "device": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
            "fs_type": null
        }, 
        {
            "action": "create format", 
            "device": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
            "fs_type": "xfs"
        }
    ], 
    "changed": true, 
    "crypts": [
        {
            "backing_device": "/dev/sda", 
            "name": "luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
            "password": "-", 
            "state": "absent"
        }, 
        {
            "backing_device": "/dev/sda1", 
            "name": "luks-1131d494-407c-46cd-b255-96712a964fd9", 
            "password": "-", 
            "state": "present"
        }
    ], 
    "leaves": [
        "/dev/sdb", 
        "/dev/sdc", 
        "/dev/sdd", 
        "/dev/sde", 
        "/dev/sdf", 
        "/dev/sdg", 
        "/dev/sdh", 
        "/dev/sdi", 
        "/dev/xvda1", 
        "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9"
    ], 
    "mounts": [
        {
            "fstype": "xfs", 
            "path": "/opt/test1", 
            "src": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
            "state": "absent"
        }, 
        {
            "dump": 0, 
            "fstype": "xfs", 
            "group": null, 
            "mode": null, 
            "opts": "defaults", 
            "owner": null, 
            "passno": 0, 
            "path": "/opt/test1", 
            "src": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
            "state": "mounted"
        }
    ], 
    "packages": [
        "cryptsetup", 
        "xfsprogs", 
        "e2fsprogs"
    ], 
    "pools": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_cipher": null, 
            "encryption_clevis_pin": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": null, 
            "encryption_tang_thumbprint": null, 
            "encryption_tang_url": null, 
            "grow_to_fill": false, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "shared": false, 
            "state": "present", 
            "type": "partition", 
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
                    "_kernel_device": "/dev/dm-0", 
                    "_mount_id": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
                    "_raw_device": "/dev/sda1", 
                    "_raw_kernel_device": "/dev/sda1", 
                    "cache_devices": [], 
                    "cache_mode": null, 
                    "cache_size": 0, 
                    "cached": false, 
                    "compression": null, 
                    "deduplication": null, 
                    "disks": [], 
                    "encryption": true, 
                    "encryption_cipher": null, 
                    "encryption_key": null, 
                    "encryption_key_size": null, 
                    "encryption_luks_version": null, 
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                    "fs_create_options": "", 
                    "fs_label": "", 
                    "fs_overwrite_existing": true, 
                    "fs_type": "xfs", 
                    "mount_check": 0, 
                    "mount_device_identifier": "uuid", 
                    "mount_group": null, 
                    "mount_mode": null, 
                    "mount_options": "defaults", 
                    "mount_passno": 0, 
                    "mount_point": "/opt/test1", 
                    "mount_user": null, 
                    "name": "test1", 
                    "raid_chunk_size": null, 
                    "raid_device_count": null, 
                    "raid_disks": [], 
                    "raid_level": null, 
                    "raid_metadata_version": null, 
                    "raid_spare_count": null, 
                    "raid_stripe_size": null, 
                    "size": "4g", 
                    "state": "present", 
                    "thin": false, 
                    "thin_pool_name": null, 
                    "thin_pool_size": null, 
                    "type": "partition", 
                    "vdo_pool_size": null
                }
            ]
        }
    ], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Saturday 14 December 2024  17:57:22 -0500 (0:00:11.022)       0:03:10.146 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Saturday 14 December 2024  17:57:22 -0500 (0:00:00.056)       0:03:10.202 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734216994.9268212, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "537beebb7e19f4750de8b7719edc3be6e665e909", 
        "ctime": 1734216994.9238212, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 263588, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0644", 
        "mtime": 1734216994.9238212, 
        "nlink": 1, 
        "path": "/etc/fstab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 1311, 
        "uid": 0, 
        "version": "606402765", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Saturday 14 December 2024  17:57:23 -0500 (0:00:00.395)       0:03:10.598 ***** 
ok: [managed-node2] => {
    "backup": "", 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  17:57:23 -0500 (0:00:00.399)       0:03:10.997 ***** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Saturday 14 December 2024  17:57:23 -0500 (0:00:00.048)       0:03:11.046 ***** 
ok: [managed-node2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format", 
                "device": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
                "fs_type": "xfs"
            }, 
            {
                "action": "destroy device", 
                "device": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
                "fs_type": null
            }, 
            {
                "action": "destroy format", 
                "device": "/dev/sda", 
                "fs_type": "luks"
            }, 
            {
                "action": "create format", 
                "device": "/dev/sda", 
                "fs_type": "disklabel"
            }, 
            {
                "action": "create device", 
                "device": "/dev/sda1", 
                "fs_type": null
            }, 
            {
                "action": "create format", 
                "device": "/dev/sda1", 
                "fs_type": "luks"
            }, 
            {
                "action": "create device", 
                "device": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
                "fs_type": null
            }, 
            {
                "action": "create format", 
                "device": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
                "fs_type": "xfs"
            }
        ], 
        "changed": true, 
        "crypts": [
            {
                "backing_device": "/dev/sda", 
                "name": "luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
                "password": "-", 
                "state": "absent"
            }, 
            {
                "backing_device": "/dev/sda1", 
                "name": "luks-1131d494-407c-46cd-b255-96712a964fd9", 
                "password": "-", 
                "state": "present"
            }
        ], 
        "failed": false, 
        "leaves": [
            "/dev/sdb", 
            "/dev/sdc", 
            "/dev/sdd", 
            "/dev/sde", 
            "/dev/sdf", 
            "/dev/sdg", 
            "/dev/sdh", 
            "/dev/sdi", 
            "/dev/xvda1", 
            "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9"
        ], 
        "mounts": [
            {
                "fstype": "xfs", 
                "path": "/opt/test1", 
                "src": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
                "state": "absent"
            }, 
            {
                "dump": 0, 
                "fstype": "xfs", 
                "group": null, 
                "mode": null, 
                "opts": "defaults", 
                "owner": null, 
                "passno": 0, 
                "path": "/opt/test1", 
                "src": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
                "state": "mounted"
            }
        ], 
        "packages": [
            "cryptsetup", 
            "xfsprogs", 
            "e2fsprogs"
        ], 
        "pools": [
            {
                "disks": [
                    "sda"
                ], 
                "encryption": false, 
                "encryption_cipher": null, 
                "encryption_clevis_pin": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": null, 
                "encryption_tang_thumbprint": null, 
                "encryption_tang_url": null, 
                "grow_to_fill": false, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "shared": false, 
                "state": "present", 
                "type": "partition", 
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
                        "_kernel_device": "/dev/dm-0", 
                        "_mount_id": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
                        "_raw_device": "/dev/sda1", 
                        "_raw_kernel_device": "/dev/sda1", 
                        "cache_devices": [], 
                        "cache_mode": null, 
                        "cache_size": 0, 
                        "cached": false, 
                        "compression": null, 
                        "deduplication": null, 
                        "disks": [], 
                        "encryption": true, 
                        "encryption_cipher": null, 
                        "encryption_key": null, 
                        "encryption_key_size": null, 
                        "encryption_luks_version": null, 
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                        "fs_create_options": "", 
                        "fs_label": "", 
                        "fs_overwrite_existing": true, 
                        "fs_type": "xfs", 
                        "mount_check": 0, 
                        "mount_device_identifier": "uuid", 
                        "mount_group": null, 
                        "mount_mode": null, 
                        "mount_options": "defaults", 
                        "mount_passno": 0, 
                        "mount_point": "/opt/test1", 
                        "mount_user": null, 
                        "name": "test1", 
                        "raid_chunk_size": null, 
                        "raid_device_count": null, 
                        "raid_disks": [], 
                        "raid_level": null, 
                        "raid_metadata_version": null, 
                        "raid_spare_count": null, 
                        "raid_stripe_size": null, 
                        "size": "4g", 
                        "state": "present", 
                        "thin": false, 
                        "thin_pool_name": null, 
                        "thin_pool_size": null, 
                        "type": "partition", 
                        "vdo_pool_size": null
                    }
                ]
            }
        ], 
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Saturday 14 December 2024  17:57:23 -0500 (0:00:00.063)       0:03:11.110 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ], 
                "encryption": false, 
                "encryption_cipher": null, 
                "encryption_clevis_pin": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": null, 
                "encryption_tang_thumbprint": null, 
                "encryption_tang_url": null, 
                "grow_to_fill": false, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "shared": false, 
                "state": "present", 
                "type": "partition", 
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
                        "_kernel_device": "/dev/dm-0", 
                        "_mount_id": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
                        "_raw_device": "/dev/sda1", 
                        "_raw_kernel_device": "/dev/sda1", 
                        "cache_devices": [], 
                        "cache_mode": null, 
                        "cache_size": 0, 
                        "cached": false, 
                        "compression": null, 
                        "deduplication": null, 
                        "disks": [], 
                        "encryption": true, 
                        "encryption_cipher": null, 
                        "encryption_key": null, 
                        "encryption_key_size": null, 
                        "encryption_luks_version": null, 
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                        "fs_create_options": "", 
                        "fs_label": "", 
                        "fs_overwrite_existing": true, 
                        "fs_type": "xfs", 
                        "mount_check": 0, 
                        "mount_device_identifier": "uuid", 
                        "mount_group": null, 
                        "mount_mode": null, 
                        "mount_options": "defaults", 
                        "mount_passno": 0, 
                        "mount_point": "/opt/test1", 
                        "mount_user": null, 
                        "name": "test1", 
                        "raid_chunk_size": null, 
                        "raid_device_count": null, 
                        "raid_disks": [], 
                        "raid_level": null, 
                        "raid_metadata_version": null, 
                        "raid_spare_count": null, 
                        "raid_stripe_size": null, 
                        "size": "4g", 
                        "state": "present", 
                        "thin": false, 
                        "thin_pool_name": null, 
                        "thin_pool_size": null, 
                        "type": "partition", 
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Saturday 14 December 2024  17:57:23 -0500 (0:00:00.058)       0:03:11.168 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Saturday 14 December 2024  17:57:23 -0500 (0:00:00.104)       0:03:11.272 ***** 
changed: [managed-node2] => (item={u'src': u'/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667', u'state': u'absent', u'fstype': u'xfs', u'path': u'/opt/test1'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": true, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "fstype": "xfs", 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
        "state": "absent"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "/dev/mapper/luks-b832bb64-70e9-412c-8abe-83daccf8b667"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Saturday 14 December 2024  17:57:24 -0500 (0:00:00.355)       0:03:11.628 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Saturday 14 December 2024  17:57:24 -0500 (0:00:00.544)       0:03:12.172 ***** 
changed: [managed-node2] => (item={u'src': u'/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": true, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
        "state": "mounted"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Saturday 14 December 2024  17:57:25 -0500 (0:00:00.419)       0:03:12.592 ***** 
skipping: [managed-node2] => (item={u'src': u'/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'})  => {
    "ansible_loop_var": "mount_info", 
    "changed": false, 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
        "state": "mounted"
    }, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Saturday 14 December 2024  17:57:25 -0500 (0:00:00.068)       0:03:12.660 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Saturday 14 December 2024  17:57:25 -0500 (0:00:00.552)       0:03:13.213 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734216999.5808191, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "dda6d914b7a43e8286ff64c03cc18f8dd7ebe5cb", 
        "ctime": 1734216996.5198205, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 917511, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0600", 
        "mtime": 1734216996.5198205, 
        "nlink": 1, 
        "path": "/etc/crypttab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": false, 
        "roth": false, 
        "rusr": true, 
        "size": 53, 
        "uid": 0, 
        "version": "606408880", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Saturday 14 December 2024  17:57:26 -0500 (0:00:00.356)       0:03:13.569 ***** 
changed: [managed-node2] => (item={u'state': u'absent', u'password': u'-', u'name': u'luks-b832bb64-70e9-412c-8abe-83daccf8b667', u'backing_device': u'/dev/sda'}) => {
    "ansible_loop_var": "entry", 
    "backup": "", 
    "changed": true, 
    "entry": {
        "backing_device": "/dev/sda", 
        "name": "luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
        "password": "-", 
        "state": "absent"
    }, 
    "found": 1
}

MSG:

1 line(s) removed
changed: [managed-node2] => (item={u'state': u'present', u'password': u'-', u'name': u'luks-1131d494-407c-46cd-b255-96712a964fd9', u'backing_device': u'/dev/sda1'}) => {
    "ansible_loop_var": "entry", 
    "backup": "", 
    "changed": true, 
    "entry": {
        "backing_device": "/dev/sda1", 
        "name": "luks-1131d494-407c-46cd-b255-96712a964fd9", 
        "password": "-", 
        "state": "present"
    }
}

MSG:

line added

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Saturday 14 December 2024  17:57:26 -0500 (0:00:00.758)       0:03:14.328 ***** 
ok: [managed-node2]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:233
Saturday 14 December 2024  17:57:27 -0500 (0:00:00.774)       0:03:15.102 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node2

TASK [Print out pool information] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Saturday 14 December 2024  17:57:27 -0500 (0:00:00.118)       0:03:15.221 ***** 
ok: [managed-node2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_cipher": null, 
            "encryption_clevis_pin": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": null, 
            "encryption_tang_thumbprint": null, 
            "encryption_tang_url": null, 
            "grow_to_fill": false, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "shared": false, 
            "state": "present", 
            "type": "partition", 
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
                    "_kernel_device": "/dev/dm-0", 
                    "_mount_id": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
                    "_raw_device": "/dev/sda1", 
                    "_raw_kernel_device": "/dev/sda1", 
                    "cache_devices": [], 
                    "cache_mode": null, 
                    "cache_size": 0, 
                    "cached": false, 
                    "compression": null, 
                    "deduplication": null, 
                    "disks": [], 
                    "encryption": true, 
                    "encryption_cipher": null, 
                    "encryption_key": null, 
                    "encryption_key_size": null, 
                    "encryption_luks_version": null, 
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                    "fs_create_options": "", 
                    "fs_label": "", 
                    "fs_overwrite_existing": true, 
                    "fs_type": "xfs", 
                    "mount_check": 0, 
                    "mount_device_identifier": "uuid", 
                    "mount_group": null, 
                    "mount_mode": null, 
                    "mount_options": "defaults", 
                    "mount_passno": 0, 
                    "mount_point": "/opt/test1", 
                    "mount_user": null, 
                    "name": "test1", 
                    "raid_chunk_size": null, 
                    "raid_device_count": null, 
                    "raid_disks": [], 
                    "raid_level": null, 
                    "raid_metadata_version": null, 
                    "raid_spare_count": null, 
                    "raid_stripe_size": null, 
                    "size": "4g", 
                    "state": "present", 
                    "thin": false, 
                    "thin_pool_name": null, 
                    "thin_pool_size": null, 
                    "type": "partition", 
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Saturday 14 December 2024  17:57:27 -0500 (0:00:00.100)       0:03:15.321 ***** 
skipping: [managed-node2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Saturday 14 December 2024  17:57:27 -0500 (0:00:00.036)       0:03:15.357 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "info": {
        "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9": {
            "fstype": "xfs", 
            "label": "", 
            "mountpoint": "/opt/test1", 
            "name": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
            "size": "10G", 
            "type": "crypt", 
            "uuid": "9d03abe3-28c6-473e-91a0-aabd0b9ce08f"
        }, 
        "/dev/sda": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sda", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sda1": {
            "fstype": "crypto_LUKS", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sda1", 
            "size": "10G", 
            "type": "partition", 
            "uuid": "1131d494-407c-46cd-b255-96712a964fd9"
        }, 
        "/dev/sdb": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdb", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdc": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdc", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdd": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdd", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sde": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sde", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdf": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdf", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdg": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdg", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdh": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdh", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdi": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdi", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/xvda", 
            "size": "250G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda1": {
            "fstype": "ext4", 
            "label": "", 
            "mountpoint": "/", 
            "name": "/dev/xvda1", 
            "size": "250G", 
            "type": "partition", 
            "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Saturday 14 December 2024  17:57:28 -0500 (0:00:00.319)       0:03:15.676 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/fstab"
    ], 
    "delta": "0:00:00.002926", 
    "end": "2024-12-14 17:57:28.549768", 
    "rc": 0, 
    "start": "2024-12-14 17:57:28.546842"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Thu Jun 20 10:23:46 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 /                       ext4    defaults        1 1
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9 /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Saturday 14 December 2024  17:57:28 -0500 (0:00:00.322)       0:03:15.999 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/crypttab"
    ], 
    "delta": "0:00:00.003413", 
    "end": "2024-12-14 17:57:28.872566", 
    "failed_when_result": false, 
    "rc": 0, 
    "start": "2024-12-14 17:57:28.869153"
}

STDOUT:

luks-1131d494-407c-46cd-b255-96712a964fd9 /dev/sda1 -

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Saturday 14 December 2024  17:57:28 -0500 (0:00:00.316)       0:03:16.315 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node2

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.081)       0:03:16.396 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members", 
            "volumes"
        ]
    }, 
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.036)       0:03:16.433 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.036)       0:03:16.469 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.035)       0:03:16.504 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node2

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.086)       0:03:16.591 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.036)       0:03:16.627 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.037)       0:03:16.665 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.034)       0:03:16.700 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.081)       0:03:16.781 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.035)       0:03:16.817 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.035)       0:03:16.852 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.035)       0:03:16.888 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.037)       0:03:16.925 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.038)       0:03:16.964 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.8.171 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.228)       0:03:17.193 ***** 

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.032)       0:03:17.226 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node2

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.076)       0:03:17.302 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.036)       0:03:17.338 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Saturday 14 December 2024  17:57:29 -0500 (0:00:00.034)       0:03:17.373 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.035)       0:03:17.409 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.036)       0:03:17.445 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.035)       0:03:17.480 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.038)       0:03:17.518 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.035)       0:03:17.554 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.035)       0:03:17.589 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.036)       0:03:17.626 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.037)       0:03:17.664 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null, 
        "storage_test_md_chunk_size_re": null, 
        "storage_test_md_metadata_version_re": null, 
        "storage_test_md_spare_devices_re": null
    }, 
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.037)       0:03:17.702 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.080)       0:03:17.782 ***** 
skipping: [managed-node2] => (item={u'_raw_device': u'/dev/sda1', u'raid_metadata_version': None, u'mount_device_identifier': u'uuid', u'fs_type': u'xfs', u'mount_options': u'defaults', u'_device': u'/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9', u'size': u'4g', u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'_kernel_device': u'/dev/dm-0', u'encryption': True, u'raid_level': None, u'name': u'test1', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'type': u'partition', u'encryption_key_size': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'mount_passno': 0, u'_mount_id': u'/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9', u'mount_user': None, u'raid_spare_count': None, u'raid_disks': [], u'_raw_kernel_device': u'/dev/sda1', u'cache_mode': None, u'cache_devices': [], u'deduplication': None, u'mount_group': None, u'thin_pool_size': None, u'disks': [], u'cached': False, u'thin': False, u'mount_check': 0, u'cache_size': 0, u'raid_chunk_size': None, u'fs_overwrite_existing': True, u'fs_create_options': u''})  => {
    "ansible_loop_var": "storage_test_lvmraid_volume", 
    "changed": false, 
    "skip_reason": "Conditional result was False", 
    "storage_test_lvmraid_volume": {
        "_device": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
        "_kernel_device": "/dev/dm-0", 
        "_mount_id": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
        "_raw_device": "/dev/sda1", 
        "_raw_kernel_device": "/dev/sda1", 
        "cache_devices": [], 
        "cache_mode": null, 
        "cache_size": 0, 
        "cached": false, 
        "compression": null, 
        "deduplication": null, 
        "disks": [], 
        "encryption": true, 
        "encryption_cipher": null, 
        "encryption_key": null, 
        "encryption_key_size": null, 
        "encryption_luks_version": null, 
        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
        "fs_create_options": "", 
        "fs_label": "", 
        "fs_overwrite_existing": true, 
        "fs_type": "xfs", 
        "mount_check": 0, 
        "mount_device_identifier": "uuid", 
        "mount_group": null, 
        "mount_mode": null, 
        "mount_options": "defaults", 
        "mount_passno": 0, 
        "mount_point": "/opt/test1", 
        "mount_user": null, 
        "name": "test1", 
        "raid_chunk_size": null, 
        "raid_device_count": null, 
        "raid_disks": [], 
        "raid_level": null, 
        "raid_metadata_version": null, 
        "raid_spare_count": null, 
        "raid_stripe_size": null, 
        "size": "4g", 
        "state": "present", 
        "thin": false, 
        "thin_pool_name": null, 
        "thin_pool_size": null, 
        "type": "partition", 
        "vdo_pool_size": null
    }
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.052)       0:03:17.835 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.081)       0:03:17.916 ***** 
skipping: [managed-node2] => (item={u'_raw_device': u'/dev/sda1', u'raid_metadata_version': None, u'mount_device_identifier': u'uuid', u'fs_type': u'xfs', u'mount_options': u'defaults', u'_device': u'/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9', u'size': u'4g', u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'_kernel_device': u'/dev/dm-0', u'encryption': True, u'raid_level': None, u'name': u'test1', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'type': u'partition', u'encryption_key_size': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'mount_passno': 0, u'_mount_id': u'/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9', u'mount_user': None, u'raid_spare_count': None, u'raid_disks': [], u'_raw_kernel_device': u'/dev/sda1', u'cache_mode': None, u'cache_devices': [], u'deduplication': None, u'mount_group': None, u'thin_pool_size': None, u'disks': [], u'cached': False, u'thin': False, u'mount_check': 0, u'cache_size': 0, u'raid_chunk_size': None, u'fs_overwrite_existing': True, u'fs_create_options': u''})  => {
    "ansible_loop_var": "storage_test_thin_volume", 
    "changed": false, 
    "skip_reason": "Conditional result was False", 
    "storage_test_thin_volume": {
        "_device": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
        "_kernel_device": "/dev/dm-0", 
        "_mount_id": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
        "_raw_device": "/dev/sda1", 
        "_raw_kernel_device": "/dev/sda1", 
        "cache_devices": [], 
        "cache_mode": null, 
        "cache_size": 0, 
        "cached": false, 
        "compression": null, 
        "deduplication": null, 
        "disks": [], 
        "encryption": true, 
        "encryption_cipher": null, 
        "encryption_key": null, 
        "encryption_key_size": null, 
        "encryption_luks_version": null, 
        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
        "fs_create_options": "", 
        "fs_label": "", 
        "fs_overwrite_existing": true, 
        "fs_type": "xfs", 
        "mount_check": 0, 
        "mount_device_identifier": "uuid", 
        "mount_group": null, 
        "mount_mode": null, 
        "mount_options": "defaults", 
        "mount_passno": 0, 
        "mount_point": "/opt/test1", 
        "mount_user": null, 
        "name": "test1", 
        "raid_chunk_size": null, 
        "raid_device_count": null, 
        "raid_disks": [], 
        "raid_level": null, 
        "raid_metadata_version": null, 
        "raid_spare_count": null, 
        "raid_stripe_size": null, 
        "size": "4g", 
        "state": "present", 
        "thin": false, 
        "thin_pool_name": null, 
        "thin_pool_size": null, 
        "type": "partition", 
        "vdo_pool_size": null
    }
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.051)       0:03:17.968 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node2

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.097)       0:03:18.065 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0", 
        "_storage_test_expected_crypttab_key_file": "-"
    }, 
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.043)       0:03:18.109 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.039)       0:03:18.148 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.037)       0:03:18.186 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null, 
        "_storage_test_crypttab_key_file": null
    }, 
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.036)       0:03:18.222 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.084)       0:03:18.307 ***** 
skipping: [managed-node2] => (item={u'_raw_device': u'/dev/sda1', u'raid_metadata_version': None, u'mount_device_identifier': u'uuid', u'fs_type': u'xfs', u'mount_options': u'defaults', u'_device': u'/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9', u'size': u'4g', u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'_kernel_device': u'/dev/dm-0', u'encryption': True, u'raid_level': None, u'name': u'test1', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'type': u'partition', u'encryption_key_size': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'mount_passno': 0, u'_mount_id': u'/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9', u'mount_user': None, u'raid_spare_count': None, u'raid_disks': [], u'_raw_kernel_device': u'/dev/sda1', u'cache_mode': None, u'cache_devices': [], u'deduplication': None, u'mount_group': None, u'thin_pool_size': None, u'disks': [], u'cached': False, u'thin': False, u'mount_check': 0, u'cache_size': 0, u'raid_chunk_size': None, u'fs_overwrite_existing': True, u'fs_create_options': u''})  => {
    "ansible_loop_var": "storage_test_vdo_volume", 
    "changed": false, 
    "skip_reason": "Conditional result was False", 
    "storage_test_vdo_volume": {
        "_device": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
        "_kernel_device": "/dev/dm-0", 
        "_mount_id": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
        "_raw_device": "/dev/sda1", 
        "_raw_kernel_device": "/dev/sda1", 
        "cache_devices": [], 
        "cache_mode": null, 
        "cache_size": 0, 
        "cached": false, 
        "compression": null, 
        "deduplication": null, 
        "disks": [], 
        "encryption": true, 
        "encryption_cipher": null, 
        "encryption_key": null, 
        "encryption_key_size": null, 
        "encryption_luks_version": null, 
        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
        "fs_create_options": "", 
        "fs_label": "", 
        "fs_overwrite_existing": true, 
        "fs_type": "xfs", 
        "mount_check": 0, 
        "mount_device_identifier": "uuid", 
        "mount_group": null, 
        "mount_mode": null, 
        "mount_options": "defaults", 
        "mount_passno": 0, 
        "mount_point": "/opt/test1", 
        "mount_user": null, 
        "name": "test1", 
        "raid_chunk_size": null, 
        "raid_device_count": null, 
        "raid_disks": [], 
        "raid_level": null, 
        "raid_metadata_version": null, 
        "raid_spare_count": null, 
        "raid_stripe_size": null, 
        "size": "4g", 
        "state": "present", 
        "thin": false, 
        "thin_pool_name": null, 
        "thin_pool_size": null, 
        "type": "partition", 
        "vdo_pool_size": null
    }
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Saturday 14 December 2024  17:57:30 -0500 (0:00:00.050)       0:03:18.357 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node2

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Saturday 14 December 2024  17:57:31 -0500 (0:00:00.091)       0:03:18.449 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Saturday 14 December 2024  17:57:31 -0500 (0:00:00.037)       0:03:18.487 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Saturday 14 December 2024  17:57:31 -0500 (0:00:00.036)       0:03:18.523 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Saturday 14 December 2024  17:57:31 -0500 (0:00:00.035)       0:03:18.558 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Saturday 14 December 2024  17:57:31 -0500 (0:00:00.036)       0:03:18.595 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Saturday 14 December 2024  17:57:31 -0500 (0:00:00.034)       0:03:18.630 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    }, 
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Saturday 14 December 2024  17:57:31 -0500 (0:00:00.035)       0:03:18.665 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null, 
        "_storage_test_expected_pv_count": null, 
        "_storage_test_expected_pv_type": null, 
        "_storage_test_pool_pvs": [], 
        "_storage_test_pool_pvs_lvm": []
    }, 
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Saturday 14 December 2024  17:57:31 -0500 (0:00:00.039)       0:03:18.704 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node2

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Saturday 14 December 2024  17:57:31 -0500 (0:00:00.071)       0:03:18.776 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true, 
        "_storage_volume_tests": [
            "mount", 
            "fstab", 
            "fs", 
            "device", 
            "encryption", 
            "md", 
            "size", 
            "cache"
        ]
    }, 
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Saturday 14 December 2024  17:57:31 -0500 (0:00:00.042)       0:03:18.819 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node2

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Saturday 14 December 2024  17:57:31 -0500 (0:00:00.242)       0:03:19.061 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9"
    }, 
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Saturday 14 December 2024  17:57:31 -0500 (0:00:00.049)       0:03:19.110 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1", 
        "storage_test_swap_expected_matches": "0"
    }, 
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Saturday 14 December 2024  17:57:31 -0500 (0:00:00.045)       0:03:19.156 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Saturday 14 December 2024  17:57:31 -0500 (0:00:00.037)       0:03:19.194 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Saturday 14 December 2024  17:57:31 -0500 (0:00:00.045)       0:03:19.240 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Saturday 14 December 2024  17:57:31 -0500 (0:00:00.037)       0:03:19.277 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Saturday 14 December 2024  17:57:31 -0500 (0:00:00.040)       0:03:19.318 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Saturday 14 December 2024  17:57:31 -0500 (0:00:00.037)       0:03:19.355 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Saturday 14 December 2024  17:57:32 -0500 (0:00:00.037)       0:03:19.393 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Saturday 14 December 2024  17:57:32 -0500 (0:00:00.037)       0:03:19.430 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Saturday 14 December 2024  17:57:32 -0500 (0:00:00.037)       0:03:19.468 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null, 
        "storage_test_mount_expected_mount_point": null, 
        "storage_test_swap_expected_matches": null, 
        "storage_test_swaps": null, 
        "storage_test_sys_node": null
    }, 
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Saturday 14 December 2024  17:57:32 -0500 (0:00:00.037)       0:03:19.506 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1", 
        "storage_test_fstab_expected_mount_options_matches": "1", 
        "storage_test_fstab_expected_mount_point_matches": "1", 
        "storage_test_fstab_id_matches": [
            "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9 "
        ], 
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ], 
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    }, 
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Saturday 14 December 2024  17:57:32 -0500 (0:00:00.066)       0:03:19.572 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Saturday 14 December 2024  17:57:32 -0500 (0:00:00.046)       0:03:19.618 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Saturday 14 December 2024  17:57:32 -0500 (0:00:00.044)       0:03:19.663 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Saturday 14 December 2024  17:57:32 -0500 (0:00:00.034)       0:03:19.698 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Saturday 14 December 2024  17:57:32 -0500 (0:00:00.044)       0:03:19.742 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null, 
        "storage_test_fstab_expected_mount_options_matches": null, 
        "storage_test_fstab_expected_mount_point_matches": null, 
        "storage_test_fstab_id_matches": null, 
        "storage_test_fstab_mount_options_matches": null, 
        "storage_test_fstab_mount_point_matches": null
    }, 
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Saturday 14 December 2024  17:57:32 -0500 (0:00:00.037)       0:03:19.779 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Saturday 14 December 2024  17:57:32 -0500 (0:00:00.051)       0:03:19.831 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Saturday 14 December 2024  17:57:32 -0500 (0:00:00.052)       0:03:19.884 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217042.5107992, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "ctime": 1734217042.5107992, 
        "dev": 5, 
        "device_type": 2049, 
        "executable": false, 
        "exists": true, 
        "gid": 6, 
        "gr_name": "disk", 
        "inode": 400365, 
        "isblk": true, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": false, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/blockdevice", 
        "mode": "0660", 
        "mtime": 1734217042.5107992, 
        "nlink": 1, 
        "path": "/dev/sda1", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": null, 
        "wgrp": true, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Saturday 14 December 2024  17:57:32 -0500 (0:00:00.326)       0:03:20.210 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Saturday 14 December 2024  17:57:32 -0500 (0:00:00.059)       0:03:20.269 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Saturday 14 December 2024  17:57:32 -0500 (0:00:00.054)       0:03:20.323 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Saturday 14 December 2024  17:57:33 -0500 (0:00:00.064)       0:03:20.388 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "st_volume_type": "partition"
    }, 
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Saturday 14 December 2024  17:57:33 -0500 (0:00:00.064)       0:03:20.453 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Saturday 14 December 2024  17:57:33 -0500 (0:00:00.049)       0:03:20.503 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Saturday 14 December 2024  17:57:33 -0500 (0:00:00.055)       0:03:20.558 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217042.6187992, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "ctime": 1734217042.6187992, 
        "dev": 5, 
        "device_type": 64768, 
        "executable": false, 
        "exists": true, 
        "gid": 6, 
        "gr_name": "disk", 
        "inode": 401434, 
        "isblk": true, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": false, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/symlink", 
        "mode": "0660", 
        "mtime": 1734217042.6187992, 
        "nlink": 1, 
        "path": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": null, 
        "wgrp": true, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Saturday 14 December 2024  17:57:33 -0500 (0:00:00.353)       0:03:20.912 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}
lsrpackages: cryptsetup

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Saturday 14 December 2024  17:57:34 -0500 (0:00:00.665)       0:03:21.578 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cryptsetup", 
        "luksDump", 
        "/dev/sda1"
    ], 
    "delta": "0:00:00.026601", 
    "end": "2024-12-14 17:57:34.499254", 
    "rc": 0, 
    "start": "2024-12-14 17:57:34.472653"
}

STDOUT:

LUKS header information for /dev/sda1

Version:       	1
Cipher name:   	aes
Cipher mode:   	xts-plain64
Hash spec:     	sha256
Payload offset:	8192
MK bits:       	512
MK digest:     	4f 34 54 48 0b 88 53 57 0e 3a a9 bf c8 61 78 ae ab 15 db 0d 
MK salt:       	47 0b 9f 80 5b e5 bc c2 76 c2 9a 03 0b 57 06 4d 
               	66 9f f8 32 36 e4 f8 99 0e e2 89 ec 03 2d 1a d7 
MK iterations: 	23108
UUID:          	1131d494-407c-46cd-b255-96712a964fd9

Key Slot 0: ENABLED
	Iterations:         	369736
	Salt:               	fa 53 2f 54 d9 c7 50 4e 51 22 e6 da 96 ee ec 0f 
	                      	87 ef 2e 7c 6a e2 35 48 0e fd 2d 2c 39 58 0a 45 
	Key material offset:	8
	AF stripes:            	4000
Key Slot 1: DISABLED
Key Slot 2: DISABLED
Key Slot 3: DISABLED
Key Slot 4: DISABLED
Key Slot 5: DISABLED
Key Slot 6: DISABLED
Key Slot 7: DISABLED

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Saturday 14 December 2024  17:57:34 -0500 (0:00:00.368)       0:03:21.946 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Saturday 14 December 2024  17:57:34 -0500 (0:00:00.060)       0:03:22.010 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Saturday 14 December 2024  17:57:34 -0500 (0:00:00.075)       0:03:22.085 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Saturday 14 December 2024  17:57:34 -0500 (0:00:00.054)       0:03:22.140 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Saturday 14 December 2024  17:57:34 -0500 (0:00:00.072)       0:03:22.212 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Saturday 14 December 2024  17:57:34 -0500 (0:00:00.050)       0:03:22.262 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Saturday 14 December 2024  17:57:34 -0500 (0:00:00.041)       0:03:22.304 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Saturday 14 December 2024  17:57:34 -0500 (0:00:00.038)       0:03:22.342 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [
            "luks-1131d494-407c-46cd-b255-96712a964fd9 /dev/sda1 -"
        ], 
        "_storage_test_expected_crypttab_entries": "1", 
        "_storage_test_expected_crypttab_key_file": "-"
    }, 
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.047)       0:03:22.390 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.045)       0:03:22.435 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.046)       0:03:22.482 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.053)       0:03:22.536 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.064)       0:03:22.600 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null, 
        "_storage_test_expected_crypttab_entries": null, 
        "_storage_test_expected_crypttab_key_file": null
    }, 
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.057)       0:03:22.658 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.059)       0:03:22.717 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.048)       0:03:22.765 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.045)       0:03:22.810 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.046)       0:03:22.857 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.042)       0:03:22.899 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.039)       0:03:22.939 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.038)       0:03:22.978 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.037)       0:03:23.015 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.036)       0:03:23.052 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.036)       0:03:23.088 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.037)       0:03:23.126 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.043)       0:03:23.170 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.056)       0:03:23.226 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.072)       0:03:23.299 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Saturday 14 December 2024  17:57:35 -0500 (0:00:00.045)       0:03:23.344 ***** 
skipping: [managed-node2] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.048)       0:03:23.393 ***** 
skipping: [managed-node2] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.045)       0:03:23.438 ***** 
skipping: [managed-node2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.042)       0:03:23.480 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.042)       0:03:23.522 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.104)       0:03:23.627 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.062)       0:03:23.690 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.051)       0:03:23.742 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.042)       0:03:23.784 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.045)       0:03:23.830 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.047)       0:03:23.878 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.047)       0:03:23.926 ***** 
skipping: [managed-node2] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.037)       0:03:23.963 ***** 
skipping: [managed-node2] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.036)       0:03:24.000 ***** 
skipping: [managed-node2] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.036)       0:03:24.037 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.036)       0:03:24.074 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.036)       0:03:24.110 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.039)       0:03:24.149 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.036)       0:03:24.185 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.036)       0:03:24.222 ***** 
ok: [managed-node2] => {
    "storage_test_actual_size": {
        "changed": false, 
        "skip_reason": "Conditional result was False", 
        "skipped": true
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.040)       0:03:24.263 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.051)       0:03:24.315 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Saturday 14 December 2024  17:57:36 -0500 (0:00:00.047)       0:03:24.363 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Saturday 14 December 2024  17:57:37 -0500 (0:00:00.053)       0:03:24.417 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Saturday 14 December 2024  17:57:37 -0500 (0:00:00.054)       0:03:24.471 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Saturday 14 December 2024  17:57:37 -0500 (0:00:00.061)       0:03:24.532 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Saturday 14 December 2024  17:57:37 -0500 (0:00:00.098)       0:03:24.630 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Saturday 14 December 2024  17:57:37 -0500 (0:00:00.119)       0:03:24.750 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Saturday 14 December 2024  17:57:37 -0500 (0:00:00.054)       0:03:24.805 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Saturday 14 December 2024  17:57:37 -0500 (0:00:00.069)       0:03:24.875 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    }, 
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Saturday 14 December 2024  17:57:37 -0500 (0:00:00.063)       0:03:24.939 ***** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Saturday 14 December 2024  17:57:37 -0500 (0:00:00.038)       0:03:24.978 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null, 
        "storage_test_crypttab": null, 
        "storage_test_fstab": null
    }, 
    "changed": false
}

TASK [Create a file] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml:12
Saturday 14 December 2024  17:57:37 -0500 (0:00:00.036)       0:03:25.014 ***** 
changed: [managed-node2] => {
    "changed": true, 
    "dest": "/opt/test1/quux", 
    "gid": 0, 
    "group": "root", 
    "mode": "0644", 
    "owner": "root", 
    "secontext": "unconfined_u:object_r:unlabeled_t:s0", 
    "size": 0, 
    "state": "file", 
    "uid": 0
}

TASK [Test for correct handling of safe_mode] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:239
Saturday 14 December 2024  17:57:38 -0500 (0:00:00.431)       0:03:25.446 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node2

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Saturday 14 December 2024  17:57:38 -0500 (0:00:00.154)       0:03:25.601 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_pools_global": [], 
        "storage_safe_mode_global": true, 
        "storage_volumes_global": []
    }, 
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Saturday 14 December 2024  17:57:38 -0500 (0:00:00.060)       0:03:25.661 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  17:57:38 -0500 (0:00:00.069)       0:03:25.730 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  17:57:38 -0500 (0:00:00.073)       0:03:25.804 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  17:57:38 -0500 (0:00:00.046)       0:03:25.850 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  17:57:38 -0500 (0:00:00.151)       0:03:26.002 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  17:57:38 -0500 (0:00:00.074)       0:03:26.077 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  17:57:38 -0500 (0:00:00.126)       0:03:26.203 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  17:57:38 -0500 (0:00:00.085)       0:03:26.289 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  17:57:39 -0500 (0:00:00.109)       0:03:26.398 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  17:57:39 -0500 (0:00:00.179)       0:03:26.578 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  17:57:40 -0500 (0:00:01.278)       0:03:27.856 ***** 
ok: [managed-node2] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ], 
            "name": "foo", 
            "type": "partition", 
            "volumes": [
                {
                    "encryption": false, 
                    "encryption_password": "yabbadabbadoo", 
                    "mount_point": "/opt/test1", 
                    "name": "test1", 
                    "size": "4g", 
                    "type": "partition"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  17:57:40 -0500 (0:00:00.067)       0:03:27.924 ***** 
ok: [managed-node2] => {
    "storage_volumes": []
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  17:57:40 -0500 (0:00:00.069)       0:03:27.993 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  17:57:44 -0500 (0:00:03.992)       0:03:31.985 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  17:57:44 -0500 (0:00:00.083)       0:03:32.069 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  17:57:44 -0500 (0:00:00.037)       0:03:32.106 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  17:57:44 -0500 (0:00:00.037)       0:03:32.144 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  17:57:44 -0500 (0:00:00.120)       0:03:32.264 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: kpartx

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  17:57:45 -0500 (0:00:00.717)       0:03:32.982 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-cryptsetup@luks\\x2db832bb64\\x2d70e9\\x2d412c\\x2d8abe\\x2d83daccf8b667.service": {
                "name": "systemd-cryptsetup@luks\\x2db832bb64\\x2d70e9\\x2d412c\\x2d8abe\\x2d83daccf8b667.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  17:57:46 -0500 (0:00:01.019)       0:03:34.002 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": [
            "systemd-cryptsetup@luks\\x2db832bb64\\x2d70e9\\x2d412c\\x2d8abe\\x2d83daccf8b667.service"
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  17:57:46 -0500 (0:00:00.067)       0:03:34.069 ***** 
changed: [managed-node2] => (item=systemd-cryptsetup@luks\x2db832bb64\x2d70e9\x2d412c\x2d8abe\x2d83daccf8b667.service) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": "systemd-cryptsetup@luks\\x2db832bb64\\x2d70e9\\x2d412c\\x2d8abe\\x2d83daccf8b667.service", 
    "name": "systemd-cryptsetup@luks\\x2db832bb64\\x2d70e9\\x2d412c\\x2d8abe\\x2d83daccf8b667.service", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "After": "dev-sda.device cryptsetup-pre.target systemd-readahead-replay.service system-systemd\\x2dcryptsetup.slice systemd-readahead-collect.service systemd-journald.socket", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "Before": "umount.target cryptsetup.target", 
        "BindsTo": "dev-sda.device", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "Conflicts": "umount.target", 
        "ControlPID": "0", 
        "DefaultDependencies": "no", 
        "Delegate": "no", 
        "Description": "Cryptography Setup for luks-b832bb64-70e9-412c-8abe-83daccf8b667", 
        "DevicePolicy": "auto", 
        "Documentation": "man:crypttab(5) man:systemd-cryptsetup-generator(8) man:systemd-cryptsetup@.service(8)", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-b832bb64-70e9-412c-8abe-83daccf8b667 /dev/sda -  ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-b832bb64-70e9-412c-8abe-83daccf8b667 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2db832bb64\\x2d70e9\\x2d412c\\x2d8abe\\x2d83daccf8b667.service", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "systemd-cryptsetup@luks\\x2db832bb64\\x2d70e9\\x2d412c\\x2d8abe\\x2d83daccf8b667.service", 
        "IgnoreOnIsolate": "yes", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "4096", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "loaded", 
        "MainPID": "0", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "systemd-cryptsetup@luks\\x2db832bb64\\x2d70e9\\x2d412c\\x2d8abe\\x2d83daccf8b667.service", 
        "NeedDaemonReload": "yes", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "yes", 
        "RequiredBy": "cryptsetup.target", 
        "Requires": "system-systemd\\x2dcryptsetup.slice", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "Slice": "system-systemd\\x2dcryptsetup.slice", 
        "SourcePath": "/etc/crypttab", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "journal", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "0", 
        "TimeoutStopUSec": "0", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "Type": "oneshot", 
        "UMask": "0022", 
        "UnitFilePreset": "disabled", 
        "UnitFileState": "bad", 
        "WantedBy": "dev-sda.device", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  17:57:47 -0500 (0:00:00.579)       0:03:34.649 ***** 
fatal: [managed-node2]: FAILED! => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [], 
    "pools": [], 
    "volumes": []
}

MSG:

cannot remove existing formatting on device 'luks-1131d494-407c-46cd-b255-96712a964fd9' in safe mode due to encryption removal

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Saturday 14 December 2024  17:57:51 -0500 (0:00:04.064)       0:03:38.714 ***** 
fatal: [managed-node2]: FAILED! => {
    "changed": false
}

MSG:

{u'_ansible_no_log': False, u'crypts': [], u'pools': [], u'leaves': [], u'changed': False, u'actions': [], u'failed': True, u'volumes': [], u'invocation': {u'module_args': {u'packages_only': False, u'disklabel_type': None, u'diskvolume_mkfs_option_map': {u'ext4': u'-F', u'ext3': u'-F', u'ext2': u'-F'}, u'safe_mode': True, u'pools': [{u'raid_metadata_version': None, u'encryption_key_size': None, u'encryption_key': None, u'encryption_luks_version': None, u'encryption_tang_url': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_tang_thumbprint': None, u'name': u'foo', u'encryption_password': None, u'encryption': False, u'disks': [u'sda'], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [{u'raid_metadata_version': None, u'mount_device_identifier': u'uuid', u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': u'4g', u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'encryption': False, u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'type': u'partition', u'encryption_key_size': 0, u'deduplication': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': u'luks1', u'raid_stripe_size': None, u'mount_passno': 0, u'mount_user': None, u'raid_spare_count': None, u'name': u'test1', u'cache_mode': None, u'raid_disks': [], u'mount_group': None, u'fs_overwrite_existing': True, u'disks': [u'sda'], u'cached': False, u'thin_pool_size': None, u'thin': False, u'mount_check': 0, u'cache_size': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}], u'shared': False, u'encryption_clevis_pin': None, u'type': u'partition', u'encryption_cipher': None, u'raid_chunk_size': None}], u'volumes': [], u'pool_defaults': {u'raid_metadata_version': None, u'encryption_cipher': None, u'encryption_key': None, u'encryption_luks_version': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_password': None, u'encryption': False, u'disks': [], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [], u'shared': False, u'type': u'lvm', u'encryption_key_size': None, u'raid_chunk_size': None}, u'volume_defaults': {u'raid_metadata_version': None, u'raid_level': None, u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': 0, u'mount_point': u'', u'compression': None, u'encryption_password': None, u'encryption': False, u'mount_device_identifier': u'uuid', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'thin_pool_name': None, u'type': u'lvm', u'encryption_key_size': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'cache_size': 0, u'raid_spare_count': None, u'cache_mode': None, u'deduplication': None, u'cached': False, u'fs_overwrite_existing': True, u'disks': [], u'thin_pool_size': None, u'thin': None, u'mount_check': 0, u'mount_passno': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}, u'use_partitions': None}}, u'mounts': [], u'packages': [], u'msg': u"cannot remove existing formatting on device 'luks-1131d494-407c-46cd-b255-96712a964fd9' in safe mode due to encryption removal"}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  17:57:51 -0500 (0:00:00.083)       0:03:38.798 ***** 
changed: [managed-node2] => (item=systemd-cryptsetup@luks\x2db832bb64\x2d70e9\x2d412c\x2d8abe\x2d83daccf8b667.service) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": "systemd-cryptsetup@luks\\x2db832bb64\\x2d70e9\\x2d412c\\x2d8abe\\x2d83daccf8b667.service", 
    "name": "systemd-cryptsetup@luks\\x2db832bb64\\x2d70e9\\x2d412c\\x2d8abe\\x2d83daccf8b667.service", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "ControlPID": "0", 
        "DefaultDependencies": "yes", 
        "Delegate": "no", 
        "Description": "systemd-cryptsetup@luks\\x2db832bb64\\x2d70e9\\x2d412c\\x2d8abe\\x2d83daccf8b667.service", 
        "DevicePolicy": "auto", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/dev/null", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "systemd-cryptsetup@luks\\x2db832bb64\\x2d70e9\\x2d412c\\x2d8abe\\x2d83daccf8b667.service", 
        "IgnoreOnIsolate": "no", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "1048576", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "masked", 
        "MainPID": "0", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "systemd-cryptsetup@luks\\x2db832bb64\\x2d70e9\\x2d412c\\x2d8abe\\x2d83daccf8b667.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "no", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "inherit", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "1min 30s", 
        "TimeoutStopUSec": "1min 30s", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "UMask": "0022", 
        "UnitFileState": "bad", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Saturday 14 December 2024  17:57:52 -0500 (0:00:00.621)       0:03:39.419 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Saturday 14 December 2024  17:57:52 -0500 (0:00:00.088)       0:03:39.508 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Saturday 14 December 2024  17:57:52 -0500 (0:00:00.071)       0:03:39.579 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Stat the file] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:11
Saturday 14 December 2024  17:57:52 -0500 (0:00:00.053)       0:03:39.633 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217057.973792, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", 
        "ctime": 1734217057.973792, 
        "dev": 64768, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 67, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/x-empty", 
        "mode": "0644", 
        "mtime": 1734217057.973792, 
        "nlink": 1, 
        "path": "/opt/test1/quux", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": "1730704867", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Assert file presence] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:16
Saturday 14 December 2024  17:57:52 -0500 (0:00:00.352)       0:03:39.986 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Remove the encryption layer] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:263
Saturday 14 December 2024  17:57:52 -0500 (0:00:00.059)       0:03:40.045 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  17:57:52 -0500 (0:00:00.214)       0:03:40.260 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  17:57:52 -0500 (0:00:00.072)       0:03:40.332 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  17:57:52 -0500 (0:00:00.044)       0:03:40.377 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  17:57:53 -0500 (0:00:00.092)       0:03:40.469 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  17:57:53 -0500 (0:00:00.044)       0:03:40.513 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  17:57:53 -0500 (0:00:00.052)       0:03:40.566 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  17:57:53 -0500 (0:00:00.054)       0:03:40.620 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  17:57:53 -0500 (0:00:00.051)       0:03:40.672 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  17:57:53 -0500 (0:00:00.201)       0:03:40.873 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  17:57:54 -0500 (0:00:01.284)       0:03:42.157 ***** 
ok: [managed-node2] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ], 
            "name": "foo", 
            "type": "partition", 
            "volumes": [
                {
                    "encryption": false, 
                    "encryption_password": "yabbadabbadoo", 
                    "mount_point": "/opt/test1", 
                    "name": "test1", 
                    "size": "4g", 
                    "type": "partition"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  17:57:54 -0500 (0:00:00.044)       0:03:42.202 ***** 
ok: [managed-node2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  17:57:54 -0500 (0:00:00.040)       0:03:42.243 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  17:57:58 -0500 (0:00:03.878)       0:03:46.121 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  17:57:58 -0500 (0:00:00.077)       0:03:46.199 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  17:57:58 -0500 (0:00:00.034)       0:03:46.234 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  17:57:58 -0500 (0:00:00.038)       0:03:46.273 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  17:57:58 -0500 (0:00:00.034)       0:03:46.307 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: kpartx

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  17:57:59 -0500 (0:00:00.632)       0:03:46.940 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service": {
                "name": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  17:58:00 -0500 (0:00:00.997)       0:03:47.938 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": [
            "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service"
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  17:58:00 -0500 (0:00:00.114)       0:03:48.052 ***** 
changed: [managed-node2] => (item=systemd-cryptsetup@luks\x2d1131d494\x2d407c\x2d46cd\x2db255\x2d96712a964fd9.service) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
    "name": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "After": "dev-sda1.device systemd-journald.socket systemd-readahead-replay.service system-systemd\\x2dcryptsetup.slice systemd-readahead-collect.service cryptsetup-pre.target", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "Before": "cryptsetup.target umount.target", 
        "BindsTo": "dev-sda1.device", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "Conflicts": "umount.target", 
        "ControlPID": "0", 
        "DefaultDependencies": "no", 
        "Delegate": "no", 
        "Description": "Cryptography Setup for luks-1131d494-407c-46cd-b255-96712a964fd9", 
        "DevicePolicy": "auto", 
        "Documentation": "man:crypttab(5) man:systemd-cryptsetup-generator(8) man:systemd-cryptsetup@.service(8)", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-1131d494-407c-46cd-b255-96712a964fd9 /dev/sda1 -  ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-1131d494-407c-46cd-b255-96712a964fd9 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
        "IgnoreOnIsolate": "yes", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "4096", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "loaded", 
        "MainPID": "0", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "yes", 
        "RequiredBy": "cryptsetup.target dev-mapper-luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.device", 
        "Requires": "system-systemd\\x2dcryptsetup.slice", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "Slice": "system-systemd\\x2dcryptsetup.slice", 
        "SourcePath": "/etc/crypttab", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "journal", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "0", 
        "TimeoutStopUSec": "0", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "Type": "oneshot", 
        "UMask": "0022", 
        "UnitFilePreset": "disabled", 
        "UnitFileState": "bad", 
        "WantedBy": "dev-sda1.device", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  17:58:01 -0500 (0:00:00.856)       0:03:48.909 ***** 
changed: [managed-node2] => {
    "actions": [
        {
            "action": "destroy format", 
            "device": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
            "fs_type": "xfs"
        }, 
        {
            "action": "destroy device", 
            "device": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
            "fs_type": null
        }, 
        {
            "action": "destroy format", 
            "device": "/dev/sda1", 
            "fs_type": "luks"
        }, 
        {
            "action": "create format", 
            "device": "/dev/sda1", 
            "fs_type": "xfs"
        }
    ], 
    "changed": true, 
    "crypts": [
        {
            "backing_device": "/dev/sda1", 
            "name": "luks-1131d494-407c-46cd-b255-96712a964fd9", 
            "password": "-", 
            "state": "absent"
        }
    ], 
    "leaves": [
        "/dev/sda1", 
        "/dev/sdb", 
        "/dev/sdc", 
        "/dev/sdd", 
        "/dev/sde", 
        "/dev/sdf", 
        "/dev/sdg", 
        "/dev/sdh", 
        "/dev/sdi", 
        "/dev/xvda1"
    ], 
    "mounts": [
        {
            "fstype": "xfs", 
            "path": "/opt/test1", 
            "src": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
            "state": "absent"
        }, 
        {
            "dump": 0, 
            "fstype": "xfs", 
            "group": null, 
            "mode": null, 
            "opts": "defaults", 
            "owner": null, 
            "passno": 0, 
            "path": "/opt/test1", 
            "src": "UUID=c3bd390d-493f-49f4-9a69-173ef13af359", 
            "state": "mounted"
        }
    ], 
    "packages": [
        "xfsprogs", 
        "e2fsprogs"
    ], 
    "pools": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_cipher": null, 
            "encryption_clevis_pin": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": null, 
            "encryption_tang_thumbprint": null, 
            "encryption_tang_url": null, 
            "grow_to_fill": false, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "shared": false, 
            "state": "present", 
            "type": "partition", 
            "volumes": [
                {
                    "_device": "/dev/sda1", 
                    "_kernel_device": "/dev/sda1", 
                    "_mount_id": "UUID=c3bd390d-493f-49f4-9a69-173ef13af359", 
                    "_raw_device": "/dev/sda1", 
                    "_raw_kernel_device": "/dev/sda1", 
                    "cache_devices": [], 
                    "cache_mode": null, 
                    "cache_size": 0, 
                    "cached": false, 
                    "compression": null, 
                    "deduplication": null, 
                    "disks": [
                        "sda"
                    ], 
                    "encryption": false, 
                    "encryption_cipher": null, 
                    "encryption_key": null, 
                    "encryption_key_size": 0, 
                    "encryption_luks_version": "luks1", 
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                    "fs_create_options": "", 
                    "fs_label": "", 
                    "fs_overwrite_existing": true, 
                    "fs_type": "xfs", 
                    "mount_check": 0, 
                    "mount_device_identifier": "uuid", 
                    "mount_group": null, 
                    "mount_mode": null, 
                    "mount_options": "defaults", 
                    "mount_passno": 0, 
                    "mount_point": "/opt/test1", 
                    "mount_user": null, 
                    "name": "test1", 
                    "raid_chunk_size": null, 
                    "raid_device_count": null, 
                    "raid_disks": [], 
                    "raid_level": null, 
                    "raid_metadata_version": null, 
                    "raid_spare_count": null, 
                    "raid_stripe_size": null, 
                    "size": "4g", 
                    "state": "present", 
                    "thin": false, 
                    "thin_pool_name": null, 
                    "thin_pool_size": null, 
                    "type": "partition", 
                    "vdo_pool_size": null
                }
            ]
        }
    ], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Saturday 14 December 2024  17:58:05 -0500 (0:00:04.384)       0:03:53.293 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Saturday 14 December 2024  17:58:05 -0500 (0:00:00.048)       0:03:53.342 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217045.113798, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "6def9588e3a52f9da6a9612b2ed4fbb14cc5658a", 
        "ctime": 1734217045.110798, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 263588, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0644", 
        "mtime": 1734217045.110798, 
        "nlink": 1, 
        "path": "/etc/fstab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 1311, 
        "uid": 0, 
        "version": "606402765", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Saturday 14 December 2024  17:58:06 -0500 (0:00:00.386)       0:03:53.729 ***** 
ok: [managed-node2] => {
    "backup": "", 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  17:58:06 -0500 (0:00:00.397)       0:03:54.127 ***** 
changed: [managed-node2] => (item=systemd-cryptsetup@luks\x2d1131d494\x2d407c\x2d46cd\x2db255\x2d96712a964fd9.service) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
    "name": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "ControlPID": "0", 
        "DefaultDependencies": "yes", 
        "Delegate": "no", 
        "Description": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
        "DevicePolicy": "auto", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/dev/null", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
        "IgnoreOnIsolate": "no", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "1048576", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "masked", 
        "MainPID": "0", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "no", 
        "RequiredBy": "dev-mapper-luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.device cryptsetup.target", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "inherit", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "1min 30s", 
        "TimeoutStopUSec": "1min 30s", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "UMask": "0022", 
        "UnitFileState": "bad", 
        "WantedBy": "dev-sda1.device", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Saturday 14 December 2024  17:58:07 -0500 (0:00:00.585)       0:03:54.712 ***** 
ok: [managed-node2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format", 
                "device": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
                "fs_type": "xfs"
            }, 
            {
                "action": "destroy device", 
                "device": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
                "fs_type": null
            }, 
            {
                "action": "destroy format", 
                "device": "/dev/sda1", 
                "fs_type": "luks"
            }, 
            {
                "action": "create format", 
                "device": "/dev/sda1", 
                "fs_type": "xfs"
            }
        ], 
        "changed": true, 
        "crypts": [
            {
                "backing_device": "/dev/sda1", 
                "name": "luks-1131d494-407c-46cd-b255-96712a964fd9", 
                "password": "-", 
                "state": "absent"
            }
        ], 
        "failed": false, 
        "leaves": [
            "/dev/sda1", 
            "/dev/sdb", 
            "/dev/sdc", 
            "/dev/sdd", 
            "/dev/sde", 
            "/dev/sdf", 
            "/dev/sdg", 
            "/dev/sdh", 
            "/dev/sdi", 
            "/dev/xvda1"
        ], 
        "mounts": [
            {
                "fstype": "xfs", 
                "path": "/opt/test1", 
                "src": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
                "state": "absent"
            }, 
            {
                "dump": 0, 
                "fstype": "xfs", 
                "group": null, 
                "mode": null, 
                "opts": "defaults", 
                "owner": null, 
                "passno": 0, 
                "path": "/opt/test1", 
                "src": "UUID=c3bd390d-493f-49f4-9a69-173ef13af359", 
                "state": "mounted"
            }
        ], 
        "packages": [
            "xfsprogs", 
            "e2fsprogs"
        ], 
        "pools": [
            {
                "disks": [
                    "sda"
                ], 
                "encryption": false, 
                "encryption_cipher": null, 
                "encryption_clevis_pin": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": null, 
                "encryption_tang_thumbprint": null, 
                "encryption_tang_url": null, 
                "grow_to_fill": false, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "shared": false, 
                "state": "present", 
                "type": "partition", 
                "volumes": [
                    {
                        "_device": "/dev/sda1", 
                        "_kernel_device": "/dev/sda1", 
                        "_mount_id": "UUID=c3bd390d-493f-49f4-9a69-173ef13af359", 
                        "_raw_device": "/dev/sda1", 
                        "_raw_kernel_device": "/dev/sda1", 
                        "cache_devices": [], 
                        "cache_mode": null, 
                        "cache_size": 0, 
                        "cached": false, 
                        "compression": null, 
                        "deduplication": null, 
                        "disks": [
                            "sda"
                        ], 
                        "encryption": false, 
                        "encryption_cipher": null, 
                        "encryption_key": null, 
                        "encryption_key_size": 0, 
                        "encryption_luks_version": "luks1", 
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                        "fs_create_options": "", 
                        "fs_label": "", 
                        "fs_overwrite_existing": true, 
                        "fs_type": "xfs", 
                        "mount_check": 0, 
                        "mount_device_identifier": "uuid", 
                        "mount_group": null, 
                        "mount_mode": null, 
                        "mount_options": "defaults", 
                        "mount_passno": 0, 
                        "mount_point": "/opt/test1", 
                        "mount_user": null, 
                        "name": "test1", 
                        "raid_chunk_size": null, 
                        "raid_device_count": null, 
                        "raid_disks": [], 
                        "raid_level": null, 
                        "raid_metadata_version": null, 
                        "raid_spare_count": null, 
                        "raid_stripe_size": null, 
                        "size": "4g", 
                        "state": "present", 
                        "thin": false, 
                        "thin_pool_name": null, 
                        "thin_pool_size": null, 
                        "type": "partition", 
                        "vdo_pool_size": null
                    }
                ]
            }
        ], 
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Saturday 14 December 2024  17:58:07 -0500 (0:00:00.076)       0:03:54.789 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ], 
                "encryption": false, 
                "encryption_cipher": null, 
                "encryption_clevis_pin": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": null, 
                "encryption_tang_thumbprint": null, 
                "encryption_tang_url": null, 
                "grow_to_fill": false, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "shared": false, 
                "state": "present", 
                "type": "partition", 
                "volumes": [
                    {
                        "_device": "/dev/sda1", 
                        "_kernel_device": "/dev/sda1", 
                        "_mount_id": "UUID=c3bd390d-493f-49f4-9a69-173ef13af359", 
                        "_raw_device": "/dev/sda1", 
                        "_raw_kernel_device": "/dev/sda1", 
                        "cache_devices": [], 
                        "cache_mode": null, 
                        "cache_size": 0, 
                        "cached": false, 
                        "compression": null, 
                        "deduplication": null, 
                        "disks": [
                            "sda"
                        ], 
                        "encryption": false, 
                        "encryption_cipher": null, 
                        "encryption_key": null, 
                        "encryption_key_size": 0, 
                        "encryption_luks_version": "luks1", 
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                        "fs_create_options": "", 
                        "fs_label": "", 
                        "fs_overwrite_existing": true, 
                        "fs_type": "xfs", 
                        "mount_check": 0, 
                        "mount_device_identifier": "uuid", 
                        "mount_group": null, 
                        "mount_mode": null, 
                        "mount_options": "defaults", 
                        "mount_passno": 0, 
                        "mount_point": "/opt/test1", 
                        "mount_user": null, 
                        "name": "test1", 
                        "raid_chunk_size": null, 
                        "raid_device_count": null, 
                        "raid_disks": [], 
                        "raid_level": null, 
                        "raid_metadata_version": null, 
                        "raid_spare_count": null, 
                        "raid_stripe_size": null, 
                        "size": "4g", 
                        "state": "present", 
                        "thin": false, 
                        "thin_pool_name": null, 
                        "thin_pool_size": null, 
                        "type": "partition", 
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Saturday 14 December 2024  17:58:07 -0500 (0:00:00.081)       0:03:54.870 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Saturday 14 December 2024  17:58:07 -0500 (0:00:00.062)       0:03:54.933 ***** 
changed: [managed-node2] => (item={u'src': u'/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9', u'state': u'absent', u'fstype': u'xfs', u'path': u'/opt/test1'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": true, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "fstype": "xfs", 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9", 
        "state": "absent"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "/dev/mapper/luks-1131d494-407c-46cd-b255-96712a964fd9"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Saturday 14 December 2024  17:58:07 -0500 (0:00:00.382)       0:03:55.315 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Saturday 14 December 2024  17:58:08 -0500 (0:00:00.552)       0:03:55.868 ***** 
changed: [managed-node2] => (item={u'src': u'UUID=c3bd390d-493f-49f4-9a69-173ef13af359', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": true, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "UUID=c3bd390d-493f-49f4-9a69-173ef13af359", 
        "state": "mounted"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "UUID=c3bd390d-493f-49f4-9a69-173ef13af359"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Saturday 14 December 2024  17:58:08 -0500 (0:00:00.382)       0:03:56.250 ***** 
skipping: [managed-node2] => (item={u'src': u'UUID=c3bd390d-493f-49f4-9a69-173ef13af359', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'})  => {
    "ansible_loop_var": "mount_info", 
    "changed": false, 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "UUID=c3bd390d-493f-49f4-9a69-173ef13af359", 
        "state": "mounted"
    }, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Saturday 14 December 2024  17:58:08 -0500 (0:00:00.048)       0:03:56.298 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Saturday 14 December 2024  17:58:09 -0500 (0:00:00.507)       0:03:56.806 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217048.8717964, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "6f5b2ce580ba98019ef796a00d34f545b1db7387", 
        "ctime": 1734217046.8547971, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 917511, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0600", 
        "mtime": 1734217046.8537972, 
        "nlink": 1, 
        "path": "/etc/crypttab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": false, 
        "roth": false, 
        "rusr": true, 
        "size": 54, 
        "uid": 0, 
        "version": "606409034", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Saturday 14 December 2024  17:58:09 -0500 (0:00:00.395)       0:03:57.201 ***** 
changed: [managed-node2] => (item={u'state': u'absent', u'password': u'-', u'name': u'luks-1131d494-407c-46cd-b255-96712a964fd9', u'backing_device': u'/dev/sda1'}) => {
    "ansible_loop_var": "entry", 
    "backup": "", 
    "changed": true, 
    "entry": {
        "backing_device": "/dev/sda1", 
        "name": "luks-1131d494-407c-46cd-b255-96712a964fd9", 
        "password": "-", 
        "state": "absent"
    }, 
    "found": 1
}

MSG:

1 line(s) removed

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Saturday 14 December 2024  17:58:10 -0500 (0:00:00.428)       0:03:57.630 ***** 
ok: [managed-node2]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:280
Saturday 14 December 2024  17:58:11 -0500 (0:00:00.851)       0:03:58.482 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node2

TASK [Print out pool information] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Saturday 14 December 2024  17:58:11 -0500 (0:00:00.171)       0:03:58.653 ***** 
ok: [managed-node2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_cipher": null, 
            "encryption_clevis_pin": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": null, 
            "encryption_tang_thumbprint": null, 
            "encryption_tang_url": null, 
            "grow_to_fill": false, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "shared": false, 
            "state": "present", 
            "type": "partition", 
            "volumes": [
                {
                    "_device": "/dev/sda1", 
                    "_kernel_device": "/dev/sda1", 
                    "_mount_id": "UUID=c3bd390d-493f-49f4-9a69-173ef13af359", 
                    "_raw_device": "/dev/sda1", 
                    "_raw_kernel_device": "/dev/sda1", 
                    "cache_devices": [], 
                    "cache_mode": null, 
                    "cache_size": 0, 
                    "cached": false, 
                    "compression": null, 
                    "deduplication": null, 
                    "disks": [
                        "sda"
                    ], 
                    "encryption": false, 
                    "encryption_cipher": null, 
                    "encryption_key": null, 
                    "encryption_key_size": 0, 
                    "encryption_luks_version": "luks1", 
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                    "fs_create_options": "", 
                    "fs_label": "", 
                    "fs_overwrite_existing": true, 
                    "fs_type": "xfs", 
                    "mount_check": 0, 
                    "mount_device_identifier": "uuid", 
                    "mount_group": null, 
                    "mount_mode": null, 
                    "mount_options": "defaults", 
                    "mount_passno": 0, 
                    "mount_point": "/opt/test1", 
                    "mount_user": null, 
                    "name": "test1", 
                    "raid_chunk_size": null, 
                    "raid_device_count": null, 
                    "raid_disks": [], 
                    "raid_level": null, 
                    "raid_metadata_version": null, 
                    "raid_spare_count": null, 
                    "raid_stripe_size": null, 
                    "size": "4g", 
                    "state": "present", 
                    "thin": false, 
                    "thin_pool_name": null, 
                    "thin_pool_size": null, 
                    "type": "partition", 
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Saturday 14 December 2024  17:58:11 -0500 (0:00:00.104)       0:03:58.758 ***** 
skipping: [managed-node2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Saturday 14 December 2024  17:58:11 -0500 (0:00:00.059)       0:03:58.817 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "info": {
        "/dev/sda": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sda", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sda1": {
            "fstype": "xfs", 
            "label": "", 
            "mountpoint": "/opt/test1", 
            "name": "/dev/sda1", 
            "size": "10G", 
            "type": "partition", 
            "uuid": "c3bd390d-493f-49f4-9a69-173ef13af359"
        }, 
        "/dev/sdb": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdb", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdc": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdc", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdd": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdd", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sde": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sde", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdf": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdf", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdg": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdg", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdh": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdh", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdi": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdi", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/xvda", 
            "size": "250G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda1": {
            "fstype": "ext4", 
            "label": "", 
            "mountpoint": "/", 
            "name": "/dev/xvda1", 
            "size": "250G", 
            "type": "partition", 
            "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Saturday 14 December 2024  17:58:11 -0500 (0:00:00.409)       0:03:59.227 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/fstab"
    ], 
    "delta": "0:00:00.003424", 
    "end": "2024-12-14 17:58:12.162510", 
    "rc": 0, 
    "start": "2024-12-14 17:58:12.159086"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Thu Jun 20 10:23:46 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 /                       ext4    defaults        1 1
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
UUID=c3bd390d-493f-49f4-9a69-173ef13af359 /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Saturday 14 December 2024  17:58:12 -0500 (0:00:00.404)       0:03:59.631 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/crypttab"
    ], 
    "delta": "0:00:00.002782", 
    "end": "2024-12-14 17:58:12.586217", 
    "failed_when_result": false, 
    "rc": 0, 
    "start": "2024-12-14 17:58:12.583435"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Saturday 14 December 2024  17:58:12 -0500 (0:00:00.419)       0:04:00.051 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node2

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Saturday 14 December 2024  17:58:12 -0500 (0:00:00.123)       0:04:00.174 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members", 
            "volumes"
        ]
    }, 
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Saturday 14 December 2024  17:58:12 -0500 (0:00:00.053)       0:04:00.228 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Saturday 14 December 2024  17:58:12 -0500 (0:00:00.056)       0:04:00.285 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Saturday 14 December 2024  17:58:12 -0500 (0:00:00.043)       0:04:00.329 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node2

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.105)       0:04:00.435 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.040)       0:04:00.475 ***** 

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.034)       0:04:00.510 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.037)       0:04:00.548 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.037)       0:04:00.586 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.042)       0:04:00.628 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.036)       0:04:00.665 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.037)       0:04:00.702 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.037)       0:04:00.739 ***** 

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.035)       0:04:00.775 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.8.171 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.245)       0:04:01.021 ***** 

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.033)       0:04:01.054 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node2

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.074)       0:04:01.128 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.037)       0:04:01.166 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.038)       0:04:01.205 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.049)       0:04:01.254 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.049)       0:04:01.303 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.036)       0:04:01.340 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Saturday 14 December 2024  17:58:13 -0500 (0:00:00.034)       0:04:01.375 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Saturday 14 December 2024  17:58:14 -0500 (0:00:00.037)       0:04:01.412 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Saturday 14 December 2024  17:58:14 -0500 (0:00:00.036)       0:04:01.448 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Saturday 14 December 2024  17:58:14 -0500 (0:00:00.038)       0:04:01.487 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Saturday 14 December 2024  17:58:14 -0500 (0:00:00.037)       0:04:01.524 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null, 
        "storage_test_md_chunk_size_re": null, 
        "storage_test_md_metadata_version_re": null, 
        "storage_test_md_spare_devices_re": null
    }, 
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Saturday 14 December 2024  17:58:14 -0500 (0:00:00.036)       0:04:01.561 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Saturday 14 December 2024  17:58:14 -0500 (0:00:00.078)       0:04:01.639 ***** 
skipping: [managed-node2] => (item={u'_raw_device': u'/dev/sda1', u'raid_metadata_version': None, u'mount_device_identifier': u'uuid', u'fs_type': u'xfs', u'mount_options': u'defaults', u'_device': u'/dev/sda1', u'size': u'4g', u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'_kernel_device': u'/dev/sda1', u'encryption': False, u'raid_level': None, u'name': u'test1', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'type': u'partition', u'encryption_key_size': 0, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': u'luks1', u'raid_stripe_size': None, u'mount_passno': 0, u'_mount_id': u'UUID=c3bd390d-493f-49f4-9a69-173ef13af359', u'mount_user': None, u'raid_spare_count': None, u'raid_disks': [], u'_raw_kernel_device': u'/dev/sda1', u'cache_mode': None, u'cache_devices': [], u'deduplication': None, u'mount_group': None, u'thin_pool_size': None, u'disks': [u'sda'], u'cached': False, u'thin': False, u'mount_check': 0, u'cache_size': 0, u'raid_chunk_size': None, u'fs_overwrite_existing': True, u'fs_create_options': u''})  => {
    "ansible_loop_var": "storage_test_lvmraid_volume", 
    "changed": false, 
    "skip_reason": "Conditional result was False", 
    "storage_test_lvmraid_volume": {
        "_device": "/dev/sda1", 
        "_kernel_device": "/dev/sda1", 
        "_mount_id": "UUID=c3bd390d-493f-49f4-9a69-173ef13af359", 
        "_raw_device": "/dev/sda1", 
        "_raw_kernel_device": "/dev/sda1", 
        "cache_devices": [], 
        "cache_mode": null, 
        "cache_size": 0, 
        "cached": false, 
        "compression": null, 
        "deduplication": null, 
        "disks": [
            "sda"
        ], 
        "encryption": false, 
        "encryption_cipher": null, 
        "encryption_key": null, 
        "encryption_key_size": 0, 
        "encryption_luks_version": "luks1", 
        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
        "fs_create_options": "", 
        "fs_label": "", 
        "fs_overwrite_existing": true, 
        "fs_type": "xfs", 
        "mount_check": 0, 
        "mount_device_identifier": "uuid", 
        "mount_group": null, 
        "mount_mode": null, 
        "mount_options": "defaults", 
        "mount_passno": 0, 
        "mount_point": "/opt/test1", 
        "mount_user": null, 
        "name": "test1", 
        "raid_chunk_size": null, 
        "raid_device_count": null, 
        "raid_disks": [], 
        "raid_level": null, 
        "raid_metadata_version": null, 
        "raid_spare_count": null, 
        "raid_stripe_size": null, 
        "size": "4g", 
        "state": "present", 
        "thin": false, 
        "thin_pool_name": null, 
        "thin_pool_size": null, 
        "type": "partition", 
        "vdo_pool_size": null
    }
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Saturday 14 December 2024  17:58:14 -0500 (0:00:00.051)       0:04:01.691 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Saturday 14 December 2024  17:58:14 -0500 (0:00:00.078)       0:04:01.769 ***** 
skipping: [managed-node2] => (item={u'_raw_device': u'/dev/sda1', u'raid_metadata_version': None, u'mount_device_identifier': u'uuid', u'fs_type': u'xfs', u'mount_options': u'defaults', u'_device': u'/dev/sda1', u'size': u'4g', u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'_kernel_device': u'/dev/sda1', u'encryption': False, u'raid_level': None, u'name': u'test1', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'type': u'partition', u'encryption_key_size': 0, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': u'luks1', u'raid_stripe_size': None, u'mount_passno': 0, u'_mount_id': u'UUID=c3bd390d-493f-49f4-9a69-173ef13af359', u'mount_user': None, u'raid_spare_count': None, u'raid_disks': [], u'_raw_kernel_device': u'/dev/sda1', u'cache_mode': None, u'cache_devices': [], u'deduplication': None, u'mount_group': None, u'thin_pool_size': None, u'disks': [u'sda'], u'cached': False, u'thin': False, u'mount_check': 0, u'cache_size': 0, u'raid_chunk_size': None, u'fs_overwrite_existing': True, u'fs_create_options': u''})  => {
    "ansible_loop_var": "storage_test_thin_volume", 
    "changed": false, 
    "skip_reason": "Conditional result was False", 
    "storage_test_thin_volume": {
        "_device": "/dev/sda1", 
        "_kernel_device": "/dev/sda1", 
        "_mount_id": "UUID=c3bd390d-493f-49f4-9a69-173ef13af359", 
        "_raw_device": "/dev/sda1", 
        "_raw_kernel_device": "/dev/sda1", 
        "cache_devices": [], 
        "cache_mode": null, 
        "cache_size": 0, 
        "cached": false, 
        "compression": null, 
        "deduplication": null, 
        "disks": [
            "sda"
        ], 
        "encryption": false, 
        "encryption_cipher": null, 
        "encryption_key": null, 
        "encryption_key_size": 0, 
        "encryption_luks_version": "luks1", 
        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
        "fs_create_options": "", 
        "fs_label": "", 
        "fs_overwrite_existing": true, 
        "fs_type": "xfs", 
        "mount_check": 0, 
        "mount_device_identifier": "uuid", 
        "mount_group": null, 
        "mount_mode": null, 
        "mount_options": "defaults", 
        "mount_passno": 0, 
        "mount_point": "/opt/test1", 
        "mount_user": null, 
        "name": "test1", 
        "raid_chunk_size": null, 
        "raid_device_count": null, 
        "raid_disks": [], 
        "raid_level": null, 
        "raid_metadata_version": null, 
        "raid_spare_count": null, 
        "raid_stripe_size": null, 
        "size": "4g", 
        "state": "present", 
        "thin": false, 
        "thin_pool_name": null, 
        "thin_pool_size": null, 
        "type": "partition", 
        "vdo_pool_size": null
    }
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Saturday 14 December 2024  17:58:14 -0500 (0:00:00.061)       0:04:01.831 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node2

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Saturday 14 December 2024  17:58:14 -0500 (0:00:00.142)       0:04:01.973 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0", 
        "_storage_test_expected_crypttab_key_file": "-"
    }, 
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Saturday 14 December 2024  17:58:14 -0500 (0:00:00.043)       0:04:02.017 ***** 

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Saturday 14 December 2024  17:58:14 -0500 (0:00:00.033)       0:04:02.050 ***** 

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Saturday 14 December 2024  17:58:14 -0500 (0:00:00.033)       0:04:02.084 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null, 
        "_storage_test_crypttab_key_file": null
    }, 
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Saturday 14 December 2024  17:58:14 -0500 (0:00:00.039)       0:04:02.123 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Saturday 14 December 2024  17:58:14 -0500 (0:00:00.084)       0:04:02.207 ***** 
skipping: [managed-node2] => (item={u'_raw_device': u'/dev/sda1', u'raid_metadata_version': None, u'mount_device_identifier': u'uuid', u'fs_type': u'xfs', u'mount_options': u'defaults', u'_device': u'/dev/sda1', u'size': u'4g', u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'_kernel_device': u'/dev/sda1', u'encryption': False, u'raid_level': None, u'name': u'test1', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'type': u'partition', u'encryption_key_size': 0, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': u'luks1', u'raid_stripe_size': None, u'mount_passno': 0, u'_mount_id': u'UUID=c3bd390d-493f-49f4-9a69-173ef13af359', u'mount_user': None, u'raid_spare_count': None, u'raid_disks': [], u'_raw_kernel_device': u'/dev/sda1', u'cache_mode': None, u'cache_devices': [], u'deduplication': None, u'mount_group': None, u'thin_pool_size': None, u'disks': [u'sda'], u'cached': False, u'thin': False, u'mount_check': 0, u'cache_size': 0, u'raid_chunk_size': None, u'fs_overwrite_existing': True, u'fs_create_options': u''})  => {
    "ansible_loop_var": "storage_test_vdo_volume", 
    "changed": false, 
    "skip_reason": "Conditional result was False", 
    "storage_test_vdo_volume": {
        "_device": "/dev/sda1", 
        "_kernel_device": "/dev/sda1", 
        "_mount_id": "UUID=c3bd390d-493f-49f4-9a69-173ef13af359", 
        "_raw_device": "/dev/sda1", 
        "_raw_kernel_device": "/dev/sda1", 
        "cache_devices": [], 
        "cache_mode": null, 
        "cache_size": 0, 
        "cached": false, 
        "compression": null, 
        "deduplication": null, 
        "disks": [
            "sda"
        ], 
        "encryption": false, 
        "encryption_cipher": null, 
        "encryption_key": null, 
        "encryption_key_size": 0, 
        "encryption_luks_version": "luks1", 
        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
        "fs_create_options": "", 
        "fs_label": "", 
        "fs_overwrite_existing": true, 
        "fs_type": "xfs", 
        "mount_check": 0, 
        "mount_device_identifier": "uuid", 
        "mount_group": null, 
        "mount_mode": null, 
        "mount_options": "defaults", 
        "mount_passno": 0, 
        "mount_point": "/opt/test1", 
        "mount_user": null, 
        "name": "test1", 
        "raid_chunk_size": null, 
        "raid_device_count": null, 
        "raid_disks": [], 
        "raid_level": null, 
        "raid_metadata_version": null, 
        "raid_spare_count": null, 
        "raid_stripe_size": null, 
        "size": "4g", 
        "state": "present", 
        "thin": false, 
        "thin_pool_name": null, 
        "thin_pool_size": null, 
        "type": "partition", 
        "vdo_pool_size": null
    }
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Saturday 14 December 2024  17:58:14 -0500 (0:00:00.052)       0:04:02.260 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node2

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Saturday 14 December 2024  17:58:14 -0500 (0:00:00.093)       0:04:02.353 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.036)       0:04:02.390 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.036)       0:04:02.427 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.037)       0:04:02.465 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.039)       0:04:02.504 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.036)       0:04:02.541 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    }, 
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.036)       0:04:02.577 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null, 
        "_storage_test_expected_pv_count": null, 
        "_storage_test_expected_pv_type": null, 
        "_storage_test_pool_pvs": [], 
        "_storage_test_pool_pvs_lvm": []
    }, 
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.038)       0:04:02.616 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node2

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.071)       0:04:02.688 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true, 
        "_storage_volume_tests": [
            "mount", 
            "fstab", 
            "fs", 
            "device", 
            "encryption", 
            "md", 
            "size", 
            "cache"
        ]
    }, 
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.046)       0:04:02.734 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node2

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.190)       0:04:02.924 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/sda1"
    }, 
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.042)       0:04:02.967 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1", 
        "storage_test_swap_expected_matches": "0"
    }, 
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.044)       0:04:03.012 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.037)       0:04:03.049 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.046)       0:04:03.096 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.036)       0:04:03.132 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.036)       0:04:03.169 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.035)       0:04:03.204 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.037)       0:04:03.242 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.035)       0:04:03.278 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.038)       0:04:03.317 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null, 
        "storage_test_mount_expected_mount_point": null, 
        "storage_test_swap_expected_matches": null, 
        "storage_test_swaps": null, 
        "storage_test_sys_node": null
    }, 
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Saturday 14 December 2024  17:58:15 -0500 (0:00:00.037)       0:04:03.354 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1", 
        "storage_test_fstab_expected_mount_options_matches": "1", 
        "storage_test_fstab_expected_mount_point_matches": "1", 
        "storage_test_fstab_id_matches": [
            "UUID=c3bd390d-493f-49f4-9a69-173ef13af359 "
        ], 
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ], 
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    }, 
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Saturday 14 December 2024  17:58:16 -0500 (0:00:00.065)       0:04:03.420 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Saturday 14 December 2024  17:58:16 -0500 (0:00:00.044)       0:04:03.464 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Saturday 14 December 2024  17:58:16 -0500 (0:00:00.043)       0:04:03.508 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Saturday 14 December 2024  17:58:16 -0500 (0:00:00.037)       0:04:03.546 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Saturday 14 December 2024  17:58:16 -0500 (0:00:00.044)       0:04:03.590 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null, 
        "storage_test_fstab_expected_mount_options_matches": null, 
        "storage_test_fstab_expected_mount_point_matches": null, 
        "storage_test_fstab_id_matches": null, 
        "storage_test_fstab_mount_options_matches": null, 
        "storage_test_fstab_mount_point_matches": null
    }, 
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Saturday 14 December 2024  17:58:16 -0500 (0:00:00.039)       0:04:03.629 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Saturday 14 December 2024  17:58:16 -0500 (0:00:00.049)       0:04:03.679 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Saturday 14 December 2024  17:58:16 -0500 (0:00:00.051)       0:04:03.731 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217085.7457793, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "ctime": 1734217085.7457793, 
        "dev": 5, 
        "device_type": 2049, 
        "executable": false, 
        "exists": true, 
        "gid": 6, 
        "gr_name": "disk", 
        "inode": 411250, 
        "isblk": true, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": false, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/blockdevice", 
        "mode": "0660", 
        "mtime": 1734217085.7457793, 
        "nlink": 1, 
        "path": "/dev/sda1", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": null, 
        "wgrp": true, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Saturday 14 December 2024  17:58:16 -0500 (0:00:00.331)       0:04:04.062 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Saturday 14 December 2024  17:58:16 -0500 (0:00:00.046)       0:04:04.109 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Saturday 14 December 2024  17:58:16 -0500 (0:00:00.036)       0:04:04.145 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Saturday 14 December 2024  17:58:16 -0500 (0:00:00.044)       0:04:04.189 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "st_volume_type": "partition"
    }, 
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Saturday 14 December 2024  17:58:16 -0500 (0:00:00.040)       0:04:04.230 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Saturday 14 December 2024  17:58:16 -0500 (0:00:00.035)       0:04:04.265 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Saturday 14 December 2024  17:58:16 -0500 (0:00:00.047)       0:04:04.312 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Saturday 14 December 2024  17:58:16 -0500 (0:00:00.046)       0:04:04.359 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}
lsrpackages: cryptsetup

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Saturday 14 December 2024  17:58:17 -0500 (0:00:00.669)       0:04:05.029 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Saturday 14 December 2024  17:58:17 -0500 (0:00:00.036)       0:04:05.066 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Saturday 14 December 2024  17:58:17 -0500 (0:00:00.037)       0:04:05.104 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Saturday 14 December 2024  17:58:17 -0500 (0:00:00.062)       0:04:05.166 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Saturday 14 December 2024  17:58:17 -0500 (0:00:00.059)       0:04:05.226 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Saturday 14 December 2024  17:58:17 -0500 (0:00:00.057)       0:04:05.284 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Saturday 14 December 2024  17:58:17 -0500 (0:00:00.056)       0:04:05.340 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.058)       0:04:05.399 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.054)       0:04:05.453 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [], 
        "_storage_test_expected_crypttab_entries": "0", 
        "_storage_test_expected_crypttab_key_file": "-"
    }, 
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.066)       0:04:05.520 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.049)       0:04:05.570 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.047)       0:04:05.617 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.048)       0:04:05.666 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.042)       0:04:05.709 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null, 
        "_storage_test_expected_crypttab_entries": null, 
        "_storage_test_expected_crypttab_key_file": null
    }, 
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.037)       0:04:05.746 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.037)       0:04:05.783 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.036)       0:04:05.820 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.037)       0:04:05.858 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.037)       0:04:05.895 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.042)       0:04:05.938 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.071)       0:04:06.009 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.054)       0:04:06.064 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.046)       0:04:06.110 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.049)       0:04:06.159 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.048)       0:04:06.208 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.038)       0:04:06.246 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.039)       0:04:06.286 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Saturday 14 December 2024  17:58:18 -0500 (0:00:00.044)       0:04:06.330 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.058)       0:04:06.389 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.066)       0:04:06.456 ***** 
skipping: [managed-node2] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.058)       0:04:06.514 ***** 
skipping: [managed-node2] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.051)       0:04:06.566 ***** 
skipping: [managed-node2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.049)       0:04:06.616 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.057)       0:04:06.673 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.038)       0:04:06.712 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.037)       0:04:06.749 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.035)       0:04:06.785 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.036)       0:04:06.822 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.035)       0:04:06.858 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.039)       0:04:06.898 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.044)       0:04:06.942 ***** 
skipping: [managed-node2] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.053)       0:04:06.996 ***** 
skipping: [managed-node2] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.062)       0:04:07.059 ***** 
skipping: [managed-node2] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.045)       0:04:07.105 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.041)       0:04:07.147 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.048)       0:04:07.195 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.037)       0:04:07.232 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.035)       0:04:07.268 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.039)       0:04:07.307 ***** 
ok: [managed-node2] => {
    "storage_test_actual_size": {
        "changed": false, 
        "skip_reason": "Conditional result was False", 
        "skipped": true
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Saturday 14 December 2024  17:58:19 -0500 (0:00:00.053)       0:04:07.361 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Saturday 14 December 2024  17:58:20 -0500 (0:00:00.058)       0:04:07.419 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Saturday 14 December 2024  17:58:20 -0500 (0:00:00.058)       0:04:07.477 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Saturday 14 December 2024  17:58:20 -0500 (0:00:00.043)       0:04:07.521 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Saturday 14 December 2024  17:58:20 -0500 (0:00:00.046)       0:04:07.568 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Saturday 14 December 2024  17:58:20 -0500 (0:00:00.050)       0:04:07.618 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Saturday 14 December 2024  17:58:20 -0500 (0:00:00.042)       0:04:07.660 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Saturday 14 December 2024  17:58:20 -0500 (0:00:00.039)       0:04:07.699 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Saturday 14 December 2024  17:58:20 -0500 (0:00:00.039)       0:04:07.739 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Saturday 14 December 2024  17:58:20 -0500 (0:00:00.037)       0:04:07.777 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    }, 
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Saturday 14 December 2024  17:58:20 -0500 (0:00:00.037)       0:04:07.815 ***** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Saturday 14 December 2024  17:58:20 -0500 (0:00:00.033)       0:04:07.849 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null, 
        "storage_test_crypttab": null, 
        "storage_test_fstab": null
    }, 
    "changed": false
}

TASK [Create a file] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml:12
Saturday 14 December 2024  17:58:20 -0500 (0:00:00.038)       0:04:07.887 ***** 
changed: [managed-node2] => {
    "changed": true, 
    "dest": "/opt/test1/quux", 
    "gid": 0, 
    "group": "root", 
    "mode": "0644", 
    "owner": "root", 
    "secontext": "unconfined_u:object_r:unlabeled_t:s0", 
    "size": 0, 
    "state": "file", 
    "uid": 0
}

TASK [Test for correct handling of safe_mode] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:286
Saturday 14 December 2024  17:58:20 -0500 (0:00:00.352)       0:04:08.240 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node2

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Saturday 14 December 2024  17:58:21 -0500 (0:00:00.161)       0:04:08.402 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_pools_global": [], 
        "storage_safe_mode_global": true, 
        "storage_volumes_global": []
    }, 
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Saturday 14 December 2024  17:58:21 -0500 (0:00:00.058)       0:04:08.461 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  17:58:21 -0500 (0:00:00.070)       0:04:08.532 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  17:58:21 -0500 (0:00:00.125)       0:04:08.657 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  17:58:21 -0500 (0:00:00.056)       0:04:08.714 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  17:58:21 -0500 (0:00:00.119)       0:04:08.833 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  17:58:21 -0500 (0:00:00.047)       0:04:08.881 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  17:58:21 -0500 (0:00:00.047)       0:04:08.929 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  17:58:21 -0500 (0:00:00.045)       0:04:08.974 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  17:58:21 -0500 (0:00:00.039)       0:04:09.013 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  17:58:21 -0500 (0:00:00.107)       0:04:09.120 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  17:58:22 -0500 (0:00:01.235)       0:04:10.356 ***** 
ok: [managed-node2] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ], 
            "name": "foo", 
            "type": "partition", 
            "volumes": [
                {
                    "encryption": true, 
                    "encryption_password": "yabbadabbadoo", 
                    "mount_point": "/opt/test1", 
                    "name": "test1", 
                    "size": "4g", 
                    "type": "partition"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  17:58:23 -0500 (0:00:00.075)       0:04:10.432 ***** 
ok: [managed-node2] => {
    "storage_volumes": []
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  17:58:23 -0500 (0:00:00.069)       0:04:10.502 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [
        "cryptsetup"
    ], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  17:58:27 -0500 (0:00:04.140)       0:04:14.642 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  17:58:27 -0500 (0:00:00.103)       0:04:14.746 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  17:58:27 -0500 (0:00:00.052)       0:04:14.798 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  17:58:27 -0500 (0:00:00.059)       0:04:14.857 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  17:58:27 -0500 (0:00:00.055)       0:04:14.913 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed", 
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: cryptsetup kpartx

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  17:58:28 -0500 (0:00:00.728)       0:04:15.641 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service": {
                "name": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  17:58:29 -0500 (0:00:00.975)       0:04:16.617 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": [
            "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service"
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  17:58:29 -0500 (0:00:00.055)       0:04:16.673 ***** 
changed: [managed-node2] => (item=systemd-cryptsetup@luks\x2d1131d494\x2d407c\x2d46cd\x2db255\x2d96712a964fd9.service) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
    "name": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "After": "systemd-readahead-collect.service systemd-readahead-replay.service systemd-journald.socket dev-sda1.device system-systemd\\x2dcryptsetup.slice cryptsetup-pre.target", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "Before": "umount.target cryptsetup.target", 
        "BindsTo": "dev-sda1.device", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "Conflicts": "umount.target", 
        "ControlPID": "0", 
        "DefaultDependencies": "no", 
        "Delegate": "no", 
        "Description": "Cryptography Setup for luks-1131d494-407c-46cd-b255-96712a964fd9", 
        "DevicePolicy": "auto", 
        "Documentation": "man:crypttab(5) man:systemd-cryptsetup-generator(8) man:systemd-cryptsetup@.service(8)", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-1131d494-407c-46cd-b255-96712a964fd9 /dev/sda1 -  ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-1131d494-407c-46cd-b255-96712a964fd9 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
        "IgnoreOnIsolate": "yes", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "4096", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "loaded", 
        "MainPID": "0", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
        "NeedDaemonReload": "yes", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "yes", 
        "RequiredBy": "cryptsetup.target", 
        "Requires": "system-systemd\\x2dcryptsetup.slice", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "Slice": "system-systemd\\x2dcryptsetup.slice", 
        "SourcePath": "/etc/crypttab", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "journal", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "0", 
        "TimeoutStopUSec": "0", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "Type": "oneshot", 
        "UMask": "0022", 
        "UnitFilePreset": "disabled", 
        "UnitFileState": "bad", 
        "WantedBy": "dev-sda1.device", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  17:58:29 -0500 (0:00:00.494)       0:04:17.167 ***** 
fatal: [managed-node2]: FAILED! => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [], 
    "pools": [], 
    "volumes": []
}

MSG:

cannot remove existing formatting on device 'sda1' in safe mode due to adding encryption

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Saturday 14 December 2024  17:58:33 -0500 (0:00:04.156)       0:04:21.323 ***** 
fatal: [managed-node2]: FAILED! => {
    "changed": false
}

MSG:

{u'_ansible_no_log': False, u'crypts': [], u'pools': [], u'leaves': [], u'changed': False, u'actions': [], u'failed': True, u'volumes': [], u'invocation': {u'module_args': {u'packages_only': False, u'disklabel_type': None, u'diskvolume_mkfs_option_map': {u'ext4': u'-F', u'ext3': u'-F', u'ext2': u'-F'}, u'safe_mode': True, u'pools': [{u'raid_metadata_version': None, u'encryption_key_size': None, u'encryption_key': None, u'encryption_luks_version': None, u'encryption_tang_url': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_tang_thumbprint': None, u'name': u'foo', u'encryption_password': None, u'encryption': False, u'disks': [u'sda'], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [{u'raid_metadata_version': None, u'mount_device_identifier': u'uuid', u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': u'4g', u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'encryption': True, u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'type': u'partition', u'encryption_key_size': None, u'deduplication': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'mount_passno': 0, u'mount_user': None, u'raid_spare_count': None, u'name': u'test1', u'cache_mode': None, u'raid_disks': [], u'mount_group': None, u'fs_overwrite_existing': True, u'disks': [u'sda'], u'cached': False, u'thin_pool_size': None, u'thin': False, u'mount_check': 0, u'cache_size': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}], u'shared': False, u'encryption_clevis_pin': None, u'type': u'partition', u'encryption_cipher': None, u'raid_chunk_size': None}], u'volumes': [], u'pool_defaults': {u'raid_metadata_version': None, u'encryption_cipher': None, u'encryption_key': None, u'encryption_luks_version': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_password': None, u'encryption': False, u'disks': [], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [], u'shared': False, u'type': u'lvm', u'encryption_key_size': None, u'raid_chunk_size': None}, u'volume_defaults': {u'raid_metadata_version': None, u'raid_level': None, u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': 0, u'mount_point': u'', u'compression': None, u'encryption_password': None, u'encryption': False, u'mount_device_identifier': u'uuid', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'thin_pool_name': None, u'type': u'lvm', u'encryption_key_size': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'cache_size': 0, u'raid_spare_count': None, u'cache_mode': None, u'deduplication': None, u'cached': False, u'fs_overwrite_existing': True, u'disks': [], u'thin_pool_size': None, u'thin': None, u'mount_check': 0, u'mount_passno': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}, u'use_partitions': None}}, u'mounts': [], u'packages': [], u'msg': u"cannot remove existing formatting on device 'sda1' in safe mode due to adding encryption"}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  17:58:34 -0500 (0:00:00.068)       0:04:21.391 ***** 
changed: [managed-node2] => (item=systemd-cryptsetup@luks\x2d1131d494\x2d407c\x2d46cd\x2db255\x2d96712a964fd9.service) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
    "name": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "ControlPID": "0", 
        "DefaultDependencies": "yes", 
        "Delegate": "no", 
        "Description": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
        "DevicePolicy": "auto", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/dev/null", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
        "IgnoreOnIsolate": "no", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "1048576", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "masked", 
        "MainPID": "0", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "systemd-cryptsetup@luks\\x2d1131d494\\x2d407c\\x2d46cd\\x2db255\\x2d96712a964fd9.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "no", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "inherit", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "1min 30s", 
        "TimeoutStopUSec": "1min 30s", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "UMask": "0022", 
        "UnitFileState": "bad", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Saturday 14 December 2024  17:58:34 -0500 (0:00:00.592)       0:04:21.984 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Saturday 14 December 2024  17:58:34 -0500 (0:00:00.076)       0:04:22.061 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Saturday 14 December 2024  17:58:34 -0500 (0:00:00.083)       0:04:22.144 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Stat the file] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:11
Saturday 14 December 2024  17:58:34 -0500 (0:00:00.074)       0:04:22.219 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217100.7767723, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", 
        "ctime": 1734217100.7767723, 
        "dev": 2049, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 67, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/x-empty", 
        "mode": "0644", 
        "mtime": 1734217100.7767723, 
        "nlink": 1, 
        "path": "/opt/test1/quux", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": "18446744071969666291", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Assert file presence] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:16
Saturday 14 December 2024  17:58:35 -0500 (0:00:00.582)       0:04:22.802 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Create a key file] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:312
Saturday 14 December 2024  17:58:35 -0500 (0:00:00.051)       0:04:22.853 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "gid": 0, 
    "group": "root", 
    "mode": "0600", 
    "owner": "root", 
    "path": "/tmp/storage_test3FfHdzlukskey", 
    "secontext": "unconfined_u:object_r:user_tmp_t:s0", 
    "size": 0, 
    "state": "file", 
    "uid": 0
}

TASK [Write the key into the key file] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:319
Saturday 14 December 2024  17:58:36 -0500 (0:00:00.572)       0:04:23.426 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "checksum": "7a4dff3752e2baf5617c57eaac048e2b95e8af91", 
    "dest": "/tmp/storage_test3FfHdzlukskey", 
    "gid": 0, 
    "group": "root", 
    "md5sum": "4ac07b967150835c00d0865161e48744", 
    "mode": "0600", 
    "owner": "root", 
    "secontext": "unconfined_u:object_r:user_tmp_t:s0", 
    "size": 32, 
    "src": "/root/.ansible/tmp/ansible-tmp-1734217116.09-27686-111320368633416/source", 
    "state": "file", 
    "uid": 0
}

TASK [Add encryption to the volume] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:326
Saturday 14 December 2024  17:58:36 -0500 (0:00:00.821)       0:04:24.247 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  17:58:36 -0500 (0:00:00.106)       0:04:24.354 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  17:58:37 -0500 (0:00:00.088)       0:04:24.442 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  17:58:37 -0500 (0:00:00.066)       0:04:24.509 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  17:58:37 -0500 (0:00:00.131)       0:04:24.641 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  17:58:37 -0500 (0:00:00.057)       0:04:24.698 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  17:58:37 -0500 (0:00:00.067)       0:04:24.766 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  17:58:37 -0500 (0:00:00.062)       0:04:24.828 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  17:58:37 -0500 (0:00:00.053)       0:04:24.882 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  17:58:37 -0500 (0:00:00.122)       0:04:25.005 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  17:58:38 -0500 (0:00:01.299)       0:04:26.304 ***** 
ok: [managed-node2] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ], 
            "name": "foo", 
            "type": "partition", 
            "volumes": [
                {
                    "encryption": true, 
                    "encryption_key": "/tmp/storage_test3FfHdzlukskey", 
                    "mount_point": "/opt/test1", 
                    "name": "test1", 
                    "size": "4g", 
                    "type": "partition"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  17:58:38 -0500 (0:00:00.044)       0:04:26.349 ***** 
ok: [managed-node2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  17:58:39 -0500 (0:00:00.040)       0:04:26.389 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [
        "cryptsetup"
    ], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  17:58:42 -0500 (0:00:03.987)       0:04:30.377 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  17:58:43 -0500 (0:00:00.067)       0:04:30.445 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  17:58:43 -0500 (0:00:00.034)       0:04:30.479 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  17:58:43 -0500 (0:00:00.037)       0:04:30.517 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  17:58:43 -0500 (0:00:00.034)       0:04:30.551 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed", 
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: cryptsetup kpartx

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  17:58:43 -0500 (0:00:00.625)       0:04:31.176 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  17:58:44 -0500 (0:00:00.953)       0:04:32.130 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  17:58:44 -0500 (0:00:00.054)       0:04:32.185 ***** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  17:58:44 -0500 (0:00:00.036)       0:04:32.221 ***** 
changed: [managed-node2] => {
    "actions": [
        {
            "action": "destroy format", 
            "device": "/dev/sda1", 
            "fs_type": "xfs"
        }, 
        {
            "action": "create format", 
            "device": "/dev/sda1", 
            "fs_type": "luks"
        }, 
        {
            "action": "create device", 
            "device": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
            "fs_type": null
        }, 
        {
            "action": "create format", 
            "device": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
            "fs_type": "xfs"
        }
    ], 
    "changed": true, 
    "crypts": [
        {
            "backing_device": "/dev/sda1", 
            "name": "luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
            "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
            "state": "present"
        }
    ], 
    "leaves": [
        "/dev/sdb", 
        "/dev/sdc", 
        "/dev/sdd", 
        "/dev/sde", 
        "/dev/sdf", 
        "/dev/sdg", 
        "/dev/sdh", 
        "/dev/sdi", 
        "/dev/xvda1", 
        "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd"
    ], 
    "mounts": [
        {
            "fstype": "xfs", 
            "path": "/opt/test1", 
            "src": "UUID=c3bd390d-493f-49f4-9a69-173ef13af359", 
            "state": "absent"
        }, 
        {
            "dump": 0, 
            "fstype": "xfs", 
            "group": null, 
            "mode": null, 
            "opts": "defaults", 
            "owner": null, 
            "passno": 0, 
            "path": "/opt/test1", 
            "src": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
            "state": "mounted"
        }
    ], 
    "packages": [
        "cryptsetup", 
        "xfsprogs", 
        "e2fsprogs"
    ], 
    "pools": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_cipher": null, 
            "encryption_clevis_pin": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": null, 
            "encryption_tang_thumbprint": null, 
            "encryption_tang_url": null, 
            "grow_to_fill": false, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "shared": false, 
            "state": "present", 
            "type": "partition", 
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
                    "_kernel_device": "/dev/dm-0", 
                    "_mount_id": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
                    "_raw_device": "/dev/sda1", 
                    "_raw_kernel_device": "/dev/sda1", 
                    "cache_devices": [], 
                    "cache_mode": null, 
                    "cache_size": 0, 
                    "cached": false, 
                    "compression": null, 
                    "deduplication": null, 
                    "disks": [
                        "sda"
                    ], 
                    "encryption": true, 
                    "encryption_cipher": null, 
                    "encryption_key": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                    "encryption_key_size": null, 
                    "encryption_luks_version": null, 
                    "encryption_password": null, 
                    "fs_create_options": "", 
                    "fs_label": "", 
                    "fs_overwrite_existing": true, 
                    "fs_type": "xfs", 
                    "mount_check": 0, 
                    "mount_device_identifier": "uuid", 
                    "mount_group": null, 
                    "mount_mode": null, 
                    "mount_options": "defaults", 
                    "mount_passno": 0, 
                    "mount_point": "/opt/test1", 
                    "mount_user": null, 
                    "name": "test1", 
                    "raid_chunk_size": null, 
                    "raid_device_count": null, 
                    "raid_disks": [], 
                    "raid_level": null, 
                    "raid_metadata_version": null, 
                    "raid_spare_count": null, 
                    "raid_stripe_size": null, 
                    "size": "4g", 
                    "state": "present", 
                    "thin": false, 
                    "thin_pool_name": null, 
                    "thin_pool_size": null, 
                    "type": "partition", 
                    "vdo_pool_size": null
                }
            ]
        }
    ], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Saturday 14 December 2024  17:58:55 -0500 (0:00:10.346)       0:04:42.568 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Saturday 14 December 2024  17:58:55 -0500 (0:00:00.059)       0:04:42.628 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217088.7907777, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "0cb93dea5fdbf125533dc370eaec490591453c7a", 
        "ctime": 1734217088.7887778, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 263588, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0644", 
        "mtime": 1734217088.7887778, 
        "nlink": 1, 
        "path": "/etc/fstab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 1299, 
        "uid": 0, 
        "version": "606402765", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Saturday 14 December 2024  17:58:55 -0500 (0:00:00.408)       0:04:43.037 ***** 
ok: [managed-node2] => {
    "backup": "", 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  17:58:56 -0500 (0:00:00.396)       0:04:43.433 ***** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Saturday 14 December 2024  17:58:56 -0500 (0:00:00.055)       0:04:43.489 ***** 
ok: [managed-node2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format", 
                "device": "/dev/sda1", 
                "fs_type": "xfs"
            }, 
            {
                "action": "create format", 
                "device": "/dev/sda1", 
                "fs_type": "luks"
            }, 
            {
                "action": "create device", 
                "device": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
                "fs_type": null
            }, 
            {
                "action": "create format", 
                "device": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
                "fs_type": "xfs"
            }
        ], 
        "changed": true, 
        "crypts": [
            {
                "backing_device": "/dev/sda1", 
                "name": "luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
                "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                "state": "present"
            }
        ], 
        "failed": false, 
        "leaves": [
            "/dev/sdb", 
            "/dev/sdc", 
            "/dev/sdd", 
            "/dev/sde", 
            "/dev/sdf", 
            "/dev/sdg", 
            "/dev/sdh", 
            "/dev/sdi", 
            "/dev/xvda1", 
            "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd"
        ], 
        "mounts": [
            {
                "fstype": "xfs", 
                "path": "/opt/test1", 
                "src": "UUID=c3bd390d-493f-49f4-9a69-173ef13af359", 
                "state": "absent"
            }, 
            {
                "dump": 0, 
                "fstype": "xfs", 
                "group": null, 
                "mode": null, 
                "opts": "defaults", 
                "owner": null, 
                "passno": 0, 
                "path": "/opt/test1", 
                "src": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
                "state": "mounted"
            }
        ], 
        "packages": [
            "cryptsetup", 
            "xfsprogs", 
            "e2fsprogs"
        ], 
        "pools": [
            {
                "disks": [
                    "sda"
                ], 
                "encryption": false, 
                "encryption_cipher": null, 
                "encryption_clevis_pin": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": null, 
                "encryption_tang_thumbprint": null, 
                "encryption_tang_url": null, 
                "grow_to_fill": false, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "shared": false, 
                "state": "present", 
                "type": "partition", 
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
                        "_kernel_device": "/dev/dm-0", 
                        "_mount_id": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
                        "_raw_device": "/dev/sda1", 
                        "_raw_kernel_device": "/dev/sda1", 
                        "cache_devices": [], 
                        "cache_mode": null, 
                        "cache_size": 0, 
                        "cached": false, 
                        "compression": null, 
                        "deduplication": null, 
                        "disks": [
                            "sda"
                        ], 
                        "encryption": true, 
                        "encryption_cipher": null, 
                        "encryption_key": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                        "encryption_key_size": null, 
                        "encryption_luks_version": null, 
                        "encryption_password": null, 
                        "fs_create_options": "", 
                        "fs_label": "", 
                        "fs_overwrite_existing": true, 
                        "fs_type": "xfs", 
                        "mount_check": 0, 
                        "mount_device_identifier": "uuid", 
                        "mount_group": null, 
                        "mount_mode": null, 
                        "mount_options": "defaults", 
                        "mount_passno": 0, 
                        "mount_point": "/opt/test1", 
                        "mount_user": null, 
                        "name": "test1", 
                        "raid_chunk_size": null, 
                        "raid_device_count": null, 
                        "raid_disks": [], 
                        "raid_level": null, 
                        "raid_metadata_version": null, 
                        "raid_spare_count": null, 
                        "raid_stripe_size": null, 
                        "size": "4g", 
                        "state": "present", 
                        "thin": false, 
                        "thin_pool_name": null, 
                        "thin_pool_size": null, 
                        "type": "partition", 
                        "vdo_pool_size": null
                    }
                ]
            }
        ], 
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Saturday 14 December 2024  17:58:56 -0500 (0:00:00.079)       0:04:43.569 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ], 
                "encryption": false, 
                "encryption_cipher": null, 
                "encryption_clevis_pin": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": null, 
                "encryption_tang_thumbprint": null, 
                "encryption_tang_url": null, 
                "grow_to_fill": false, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "shared": false, 
                "state": "present", 
                "type": "partition", 
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
                        "_kernel_device": "/dev/dm-0", 
                        "_mount_id": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
                        "_raw_device": "/dev/sda1", 
                        "_raw_kernel_device": "/dev/sda1", 
                        "cache_devices": [], 
                        "cache_mode": null, 
                        "cache_size": 0, 
                        "cached": false, 
                        "compression": null, 
                        "deduplication": null, 
                        "disks": [
                            "sda"
                        ], 
                        "encryption": true, 
                        "encryption_cipher": null, 
                        "encryption_key": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                        "encryption_key_size": null, 
                        "encryption_luks_version": null, 
                        "encryption_password": null, 
                        "fs_create_options": "", 
                        "fs_label": "", 
                        "fs_overwrite_existing": true, 
                        "fs_type": "xfs", 
                        "mount_check": 0, 
                        "mount_device_identifier": "uuid", 
                        "mount_group": null, 
                        "mount_mode": null, 
                        "mount_options": "defaults", 
                        "mount_passno": 0, 
                        "mount_point": "/opt/test1", 
                        "mount_user": null, 
                        "name": "test1", 
                        "raid_chunk_size": null, 
                        "raid_device_count": null, 
                        "raid_disks": [], 
                        "raid_level": null, 
                        "raid_metadata_version": null, 
                        "raid_spare_count": null, 
                        "raid_stripe_size": null, 
                        "size": "4g", 
                        "state": "present", 
                        "thin": false, 
                        "thin_pool_name": null, 
                        "thin_pool_size": null, 
                        "type": "partition", 
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Saturday 14 December 2024  17:58:56 -0500 (0:00:00.072)       0:04:43.642 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Saturday 14 December 2024  17:58:56 -0500 (0:00:00.061)       0:04:43.703 ***** 
changed: [managed-node2] => (item={u'src': u'UUID=c3bd390d-493f-49f4-9a69-173ef13af359', u'state': u'absent', u'fstype': u'xfs', u'path': u'/opt/test1'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": true, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "fstype": "xfs", 
        "path": "/opt/test1", 
        "src": "UUID=c3bd390d-493f-49f4-9a69-173ef13af359", 
        "state": "absent"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "UUID=c3bd390d-493f-49f4-9a69-173ef13af359"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Saturday 14 December 2024  17:58:56 -0500 (0:00:00.446)       0:04:44.150 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Saturday 14 December 2024  17:58:57 -0500 (0:00:00.532)       0:04:44.682 ***** 
changed: [managed-node2] => (item={u'src': u'/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": true, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
        "state": "mounted"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Saturday 14 December 2024  17:58:57 -0500 (0:00:00.438)       0:04:45.120 ***** 
skipping: [managed-node2] => (item={u'src': u'/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'})  => {
    "ansible_loop_var": "mount_info", 
    "changed": false, 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
        "state": "mounted"
    }, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Saturday 14 December 2024  17:58:57 -0500 (0:00:00.075)       0:04:45.195 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Saturday 14 December 2024  17:58:58 -0500 (0:00:00.512)       0:04:45.708 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217092.584776, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", 
        "ctime": 1734217090.1467772, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 917510, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/x-empty", 
        "mode": "0600", 
        "mtime": 1734217090.1467772, 
        "nlink": 1, 
        "path": "/etc/crypttab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": false, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": "606409220", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Saturday 14 December 2024  17:58:58 -0500 (0:00:00.352)       0:04:46.061 ***** 
changed: [managed-node2] => (item={u'state': u'present', u'password': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'name': u'luks-b6506dcf-75d2-4401-8040-0091a37b6dfd', u'backing_device': u'/dev/sda1'}) => {
    "ansible_loop_var": "entry", 
    "backup": "", 
    "changed": true, 
    "entry": {
        "backing_device": "/dev/sda1", 
        "name": "luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
        "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
        "state": "present"
    }
}

MSG:

line added

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Saturday 14 December 2024  17:58:59 -0500 (0:00:00.415)       0:04:46.477 ***** 
ok: [managed-node2]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:343
Saturday 14 December 2024  17:58:59 -0500 (0:00:00.749)       0:04:47.227 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node2

TASK [Print out pool information] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Saturday 14 December 2024  17:58:59 -0500 (0:00:00.060)       0:04:47.287 ***** 
ok: [managed-node2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_cipher": null, 
            "encryption_clevis_pin": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": null, 
            "encryption_tang_thumbprint": null, 
            "encryption_tang_url": null, 
            "grow_to_fill": false, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "shared": false, 
            "state": "present", 
            "type": "partition", 
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
                    "_kernel_device": "/dev/dm-0", 
                    "_mount_id": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
                    "_raw_device": "/dev/sda1", 
                    "_raw_kernel_device": "/dev/sda1", 
                    "cache_devices": [], 
                    "cache_mode": null, 
                    "cache_size": 0, 
                    "cached": false, 
                    "compression": null, 
                    "deduplication": null, 
                    "disks": [
                        "sda"
                    ], 
                    "encryption": true, 
                    "encryption_cipher": null, 
                    "encryption_key": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                    "encryption_key_size": null, 
                    "encryption_luks_version": null, 
                    "encryption_password": null, 
                    "fs_create_options": "", 
                    "fs_label": "", 
                    "fs_overwrite_existing": true, 
                    "fs_type": "xfs", 
                    "mount_check": 0, 
                    "mount_device_identifier": "uuid", 
                    "mount_group": null, 
                    "mount_mode": null, 
                    "mount_options": "defaults", 
                    "mount_passno": 0, 
                    "mount_point": "/opt/test1", 
                    "mount_user": null, 
                    "name": "test1", 
                    "raid_chunk_size": null, 
                    "raid_device_count": null, 
                    "raid_disks": [], 
                    "raid_level": null, 
                    "raid_metadata_version": null, 
                    "raid_spare_count": null, 
                    "raid_stripe_size": null, 
                    "size": "4g", 
                    "state": "present", 
                    "thin": false, 
                    "thin_pool_name": null, 
                    "thin_pool_size": null, 
                    "type": "partition", 
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Saturday 14 December 2024  17:58:59 -0500 (0:00:00.050)       0:04:47.338 ***** 
skipping: [managed-node2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Saturday 14 December 2024  17:58:59 -0500 (0:00:00.037)       0:04:47.375 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "info": {
        "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd": {
            "fstype": "xfs", 
            "label": "", 
            "mountpoint": "/opt/test1", 
            "name": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
            "size": "10G", 
            "type": "crypt", 
            "uuid": "83602df9-d345-4644-85ad-baba82ae420a"
        }, 
        "/dev/sda": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sda", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sda1": {
            "fstype": "crypto_LUKS", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sda1", 
            "size": "10G", 
            "type": "partition", 
            "uuid": "b6506dcf-75d2-4401-8040-0091a37b6dfd"
        }, 
        "/dev/sdb": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdb", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdc": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdc", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdd": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdd", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sde": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sde", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdf": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdf", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdg": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdg", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdh": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdh", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdi": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdi", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/xvda", 
            "size": "250G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda1": {
            "fstype": "ext4", 
            "label": "", 
            "mountpoint": "/", 
            "name": "/dev/xvda1", 
            "size": "250G", 
            "type": "partition", 
            "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Saturday 14 December 2024  17:59:00 -0500 (0:00:00.342)       0:04:47.718 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/fstab"
    ], 
    "delta": "0:00:00.002958", 
    "end": "2024-12-14 17:59:00.619924", 
    "rc": 0, 
    "start": "2024-12-14 17:59:00.616966"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Thu Jun 20 10:23:46 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 /                       ext4    defaults        1 1
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Saturday 14 December 2024  17:59:00 -0500 (0:00:00.345)       0:04:48.064 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/crypttab"
    ], 
    "delta": "0:00:00.003178", 
    "end": "2024-12-14 17:59:00.966929", 
    "failed_when_result": false, 
    "rc": 0, 
    "start": "2024-12-14 17:59:00.963751"
}

STDOUT:

luks-b6506dcf-75d2-4401-8040-0091a37b6dfd /dev/sda1 VALUE_SPECIFIED_IN_NO_LOG_PARAMETER

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Saturday 14 December 2024  17:59:01 -0500 (0:00:00.348)       0:04:48.412 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node2

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Saturday 14 December 2024  17:59:01 -0500 (0:00:00.101)       0:04:48.514 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members", 
            "volumes"
        ]
    }, 
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Saturday 14 December 2024  17:59:01 -0500 (0:00:00.068)       0:04:48.582 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Saturday 14 December 2024  17:59:01 -0500 (0:00:00.071)       0:04:48.654 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Saturday 14 December 2024  17:59:01 -0500 (0:00:00.057)       0:04:48.711 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node2

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Saturday 14 December 2024  17:59:01 -0500 (0:00:00.122)       0:04:48.834 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Saturday 14 December 2024  17:59:01 -0500 (0:00:00.047)       0:04:48.881 ***** 

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Saturday 14 December 2024  17:59:01 -0500 (0:00:00.038)       0:04:48.919 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Saturday 14 December 2024  17:59:01 -0500 (0:00:00.052)       0:04:48.972 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Saturday 14 December 2024  17:59:01 -0500 (0:00:00.048)       0:04:49.020 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Saturday 14 December 2024  17:59:01 -0500 (0:00:00.043)       0:04:49.063 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Saturday 14 December 2024  17:59:01 -0500 (0:00:00.035)       0:04:49.099 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Saturday 14 December 2024  17:59:01 -0500 (0:00:00.040)       0:04:49.139 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Saturday 14 December 2024  17:59:01 -0500 (0:00:00.037)       0:04:49.177 ***** 

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Saturday 14 December 2024  17:59:01 -0500 (0:00:00.037)       0:04:49.215 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.8.171 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Saturday 14 December 2024  17:59:02 -0500 (0:00:00.273)       0:04:49.489 ***** 

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Saturday 14 December 2024  17:59:02 -0500 (0:00:00.042)       0:04:49.531 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node2

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Saturday 14 December 2024  17:59:02 -0500 (0:00:00.153)       0:04:49.685 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Saturday 14 December 2024  17:59:02 -0500 (0:00:00.056)       0:04:49.741 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Saturday 14 December 2024  17:59:02 -0500 (0:00:00.055)       0:04:49.797 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Saturday 14 December 2024  17:59:02 -0500 (0:00:00.049)       0:04:49.846 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Saturday 14 December 2024  17:59:02 -0500 (0:00:00.042)       0:04:49.889 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Saturday 14 December 2024  17:59:02 -0500 (0:00:00.052)       0:04:49.941 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Saturday 14 December 2024  17:59:02 -0500 (0:00:00.049)       0:04:49.991 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Saturday 14 December 2024  17:59:02 -0500 (0:00:00.041)       0:04:50.032 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Saturday 14 December 2024  17:59:02 -0500 (0:00:00.037)       0:04:50.070 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Saturday 14 December 2024  17:59:02 -0500 (0:00:00.037)       0:04:50.108 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Saturday 14 December 2024  17:59:02 -0500 (0:00:00.038)       0:04:50.146 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null, 
        "storage_test_md_chunk_size_re": null, 
        "storage_test_md_metadata_version_re": null, 
        "storage_test_md_spare_devices_re": null
    }, 
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Saturday 14 December 2024  17:59:02 -0500 (0:00:00.042)       0:04:50.189 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Saturday 14 December 2024  17:59:02 -0500 (0:00:00.085)       0:04:50.275 ***** 
skipping: [managed-node2] => (item={u'_raw_device': u'/dev/sda1', u'raid_metadata_version': None, u'mount_device_identifier': u'uuid', u'fs_type': u'xfs', u'mount_options': u'defaults', u'_device': u'/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd', u'size': u'4g', u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': None, u'_kernel_device': u'/dev/dm-0', u'encryption': True, u'raid_level': None, u'name': u'test1', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'type': u'partition', u'encryption_key_size': None, u'encryption_cipher': None, u'encryption_key': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'mount_passno': 0, u'_mount_id': u'/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd', u'mount_user': None, u'raid_spare_count': None, u'raid_disks': [], u'_raw_kernel_device': u'/dev/sda1', u'cache_mode': None, u'cache_devices': [], u'deduplication': None, u'mount_group': None, u'thin_pool_size': None, u'disks': [u'sda'], u'cached': False, u'thin': False, u'mount_check': 0, u'cache_size': 0, u'raid_chunk_size': None, u'fs_overwrite_existing': True, u'fs_create_options': u''})  => {
    "ansible_loop_var": "storage_test_lvmraid_volume", 
    "changed": false, 
    "skip_reason": "Conditional result was False", 
    "storage_test_lvmraid_volume": {
        "_device": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
        "_kernel_device": "/dev/dm-0", 
        "_mount_id": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
        "_raw_device": "/dev/sda1", 
        "_raw_kernel_device": "/dev/sda1", 
        "cache_devices": [], 
        "cache_mode": null, 
        "cache_size": 0, 
        "cached": false, 
        "compression": null, 
        "deduplication": null, 
        "disks": [
            "sda"
        ], 
        "encryption": true, 
        "encryption_cipher": null, 
        "encryption_key": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
        "encryption_key_size": null, 
        "encryption_luks_version": null, 
        "encryption_password": null, 
        "fs_create_options": "", 
        "fs_label": "", 
        "fs_overwrite_existing": true, 
        "fs_type": "xfs", 
        "mount_check": 0, 
        "mount_device_identifier": "uuid", 
        "mount_group": null, 
        "mount_mode": null, 
        "mount_options": "defaults", 
        "mount_passno": 0, 
        "mount_point": "/opt/test1", 
        "mount_user": null, 
        "name": "test1", 
        "raid_chunk_size": null, 
        "raid_device_count": null, 
        "raid_disks": [], 
        "raid_level": null, 
        "raid_metadata_version": null, 
        "raid_spare_count": null, 
        "raid_stripe_size": null, 
        "size": "4g", 
        "state": "present", 
        "thin": false, 
        "thin_pool_name": null, 
        "thin_pool_size": null, 
        "type": "partition", 
        "vdo_pool_size": null
    }
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Saturday 14 December 2024  17:59:02 -0500 (0:00:00.073)       0:04:50.349 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Saturday 14 December 2024  17:59:03 -0500 (0:00:00.102)       0:04:50.451 ***** 
skipping: [managed-node2] => (item={u'_raw_device': u'/dev/sda1', u'raid_metadata_version': None, u'mount_device_identifier': u'uuid', u'fs_type': u'xfs', u'mount_options': u'defaults', u'_device': u'/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd', u'size': u'4g', u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': None, u'_kernel_device': u'/dev/dm-0', u'encryption': True, u'raid_level': None, u'name': u'test1', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'type': u'partition', u'encryption_key_size': None, u'encryption_cipher': None, u'encryption_key': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'mount_passno': 0, u'_mount_id': u'/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd', u'mount_user': None, u'raid_spare_count': None, u'raid_disks': [], u'_raw_kernel_device': u'/dev/sda1', u'cache_mode': None, u'cache_devices': [], u'deduplication': None, u'mount_group': None, u'thin_pool_size': None, u'disks': [u'sda'], u'cached': False, u'thin': False, u'mount_check': 0, u'cache_size': 0, u'raid_chunk_size': None, u'fs_overwrite_existing': True, u'fs_create_options': u''})  => {
    "ansible_loop_var": "storage_test_thin_volume", 
    "changed": false, 
    "skip_reason": "Conditional result was False", 
    "storage_test_thin_volume": {
        "_device": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
        "_kernel_device": "/dev/dm-0", 
        "_mount_id": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
        "_raw_device": "/dev/sda1", 
        "_raw_kernel_device": "/dev/sda1", 
        "cache_devices": [], 
        "cache_mode": null, 
        "cache_size": 0, 
        "cached": false, 
        "compression": null, 
        "deduplication": null, 
        "disks": [
            "sda"
        ], 
        "encryption": true, 
        "encryption_cipher": null, 
        "encryption_key": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
        "encryption_key_size": null, 
        "encryption_luks_version": null, 
        "encryption_password": null, 
        "fs_create_options": "", 
        "fs_label": "", 
        "fs_overwrite_existing": true, 
        "fs_type": "xfs", 
        "mount_check": 0, 
        "mount_device_identifier": "uuid", 
        "mount_group": null, 
        "mount_mode": null, 
        "mount_options": "defaults", 
        "mount_passno": 0, 
        "mount_point": "/opt/test1", 
        "mount_user": null, 
        "name": "test1", 
        "raid_chunk_size": null, 
        "raid_device_count": null, 
        "raid_disks": [], 
        "raid_level": null, 
        "raid_metadata_version": null, 
        "raid_spare_count": null, 
        "raid_stripe_size": null, 
        "size": "4g", 
        "state": "present", 
        "thin": false, 
        "thin_pool_name": null, 
        "thin_pool_size": null, 
        "type": "partition", 
        "vdo_pool_size": null
    }
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Saturday 14 December 2024  17:59:03 -0500 (0:00:00.065)       0:04:50.517 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node2

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Saturday 14 December 2024  17:59:03 -0500 (0:00:00.086)       0:04:50.604 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0", 
        "_storage_test_expected_crypttab_key_file": "-"
    }, 
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Saturday 14 December 2024  17:59:03 -0500 (0:00:00.059)       0:04:50.663 ***** 

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Saturday 14 December 2024  17:59:03 -0500 (0:00:00.056)       0:04:50.719 ***** 

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Saturday 14 December 2024  17:59:03 -0500 (0:00:00.052)       0:04:50.772 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null, 
        "_storage_test_crypttab_key_file": null
    }, 
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Saturday 14 December 2024  17:59:03 -0500 (0:00:00.053)       0:04:50.826 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Saturday 14 December 2024  17:59:03 -0500 (0:00:00.113)       0:04:50.939 ***** 
skipping: [managed-node2] => (item={u'_raw_device': u'/dev/sda1', u'raid_metadata_version': None, u'mount_device_identifier': u'uuid', u'fs_type': u'xfs', u'mount_options': u'defaults', u'_device': u'/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd', u'size': u'4g', u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': None, u'_kernel_device': u'/dev/dm-0', u'encryption': True, u'raid_level': None, u'name': u'test1', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'type': u'partition', u'encryption_key_size': None, u'encryption_cipher': None, u'encryption_key': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'mount_passno': 0, u'_mount_id': u'/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd', u'mount_user': None, u'raid_spare_count': None, u'raid_disks': [], u'_raw_kernel_device': u'/dev/sda1', u'cache_mode': None, u'cache_devices': [], u'deduplication': None, u'mount_group': None, u'thin_pool_size': None, u'disks': [u'sda'], u'cached': False, u'thin': False, u'mount_check': 0, u'cache_size': 0, u'raid_chunk_size': None, u'fs_overwrite_existing': True, u'fs_create_options': u''})  => {
    "ansible_loop_var": "storage_test_vdo_volume", 
    "changed": false, 
    "skip_reason": "Conditional result was False", 
    "storage_test_vdo_volume": {
        "_device": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
        "_kernel_device": "/dev/dm-0", 
        "_mount_id": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
        "_raw_device": "/dev/sda1", 
        "_raw_kernel_device": "/dev/sda1", 
        "cache_devices": [], 
        "cache_mode": null, 
        "cache_size": 0, 
        "cached": false, 
        "compression": null, 
        "deduplication": null, 
        "disks": [
            "sda"
        ], 
        "encryption": true, 
        "encryption_cipher": null, 
        "encryption_key": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
        "encryption_key_size": null, 
        "encryption_luks_version": null, 
        "encryption_password": null, 
        "fs_create_options": "", 
        "fs_label": "", 
        "fs_overwrite_existing": true, 
        "fs_type": "xfs", 
        "mount_check": 0, 
        "mount_device_identifier": "uuid", 
        "mount_group": null, 
        "mount_mode": null, 
        "mount_options": "defaults", 
        "mount_passno": 0, 
        "mount_point": "/opt/test1", 
        "mount_user": null, 
        "name": "test1", 
        "raid_chunk_size": null, 
        "raid_device_count": null, 
        "raid_disks": [], 
        "raid_level": null, 
        "raid_metadata_version": null, 
        "raid_spare_count": null, 
        "raid_stripe_size": null, 
        "size": "4g", 
        "state": "present", 
        "thin": false, 
        "thin_pool_name": null, 
        "thin_pool_size": null, 
        "type": "partition", 
        "vdo_pool_size": null
    }
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Saturday 14 December 2024  17:59:03 -0500 (0:00:00.069)       0:04:51.008 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node2

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Saturday 14 December 2024  17:59:03 -0500 (0:00:00.091)       0:04:51.100 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Saturday 14 December 2024  17:59:03 -0500 (0:00:00.038)       0:04:51.139 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Saturday 14 December 2024  17:59:03 -0500 (0:00:00.036)       0:04:51.176 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Saturday 14 December 2024  17:59:03 -0500 (0:00:00.036)       0:04:51.213 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Saturday 14 December 2024  17:59:03 -0500 (0:00:00.036)       0:04:51.249 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Saturday 14 December 2024  17:59:03 -0500 (0:00:00.037)       0:04:51.286 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    }, 
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Saturday 14 December 2024  17:59:03 -0500 (0:00:00.036)       0:04:51.323 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null, 
        "_storage_test_expected_pv_count": null, 
        "_storage_test_expected_pv_type": null, 
        "_storage_test_pool_pvs": [], 
        "_storage_test_pool_pvs_lvm": []
    }, 
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Saturday 14 December 2024  17:59:03 -0500 (0:00:00.039)       0:04:51.363 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node2

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Saturday 14 December 2024  17:59:04 -0500 (0:00:00.089)       0:04:51.453 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true, 
        "_storage_volume_tests": [
            "mount", 
            "fstab", 
            "fs", 
            "device", 
            "encryption", 
            "md", 
            "size", 
            "cache"
        ]
    }, 
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Saturday 14 December 2024  17:59:04 -0500 (0:00:00.069)       0:04:51.523 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node2

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Saturday 14 December 2024  17:59:04 -0500 (0:00:00.294)       0:04:51.817 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd"
    }, 
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Saturday 14 December 2024  17:59:04 -0500 (0:00:00.057)       0:04:51.875 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1", 
        "storage_test_swap_expected_matches": "0"
    }, 
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Saturday 14 December 2024  17:59:04 -0500 (0:00:00.055)       0:04:51.930 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Saturday 14 December 2024  17:59:04 -0500 (0:00:00.047)       0:04:51.978 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Saturday 14 December 2024  17:59:04 -0500 (0:00:00.052)       0:04:52.030 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Saturday 14 December 2024  17:59:04 -0500 (0:00:00.037)       0:04:52.068 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Saturday 14 December 2024  17:59:04 -0500 (0:00:00.037)       0:04:52.105 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Saturday 14 December 2024  17:59:04 -0500 (0:00:00.045)       0:04:52.151 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Saturday 14 December 2024  17:59:04 -0500 (0:00:00.055)       0:04:52.207 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Saturday 14 December 2024  17:59:04 -0500 (0:00:00.052)       0:04:52.260 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Saturday 14 December 2024  17:59:04 -0500 (0:00:00.047)       0:04:52.307 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null, 
        "storage_test_mount_expected_mount_point": null, 
        "storage_test_swap_expected_matches": null, 
        "storage_test_swaps": null, 
        "storage_test_sys_node": null
    }, 
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Saturday 14 December 2024  17:59:04 -0500 (0:00:00.047)       0:04:52.355 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1", 
        "storage_test_fstab_expected_mount_options_matches": "1", 
        "storage_test_fstab_expected_mount_point_matches": "1", 
        "storage_test_fstab_id_matches": [
            "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd "
        ], 
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ], 
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    }, 
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Saturday 14 December 2024  17:59:05 -0500 (0:00:00.071)       0:04:52.426 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Saturday 14 December 2024  17:59:05 -0500 (0:00:00.115)       0:04:52.542 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Saturday 14 December 2024  17:59:05 -0500 (0:00:00.061)       0:04:52.604 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Saturday 14 December 2024  17:59:05 -0500 (0:00:00.050)       0:04:52.654 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Saturday 14 December 2024  17:59:05 -0500 (0:00:00.058)       0:04:52.713 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null, 
        "storage_test_fstab_expected_mount_options_matches": null, 
        "storage_test_fstab_expected_mount_point_matches": null, 
        "storage_test_fstab_id_matches": null, 
        "storage_test_fstab_mount_options_matches": null, 
        "storage_test_fstab_mount_point_matches": null
    }, 
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Saturday 14 December 2024  17:59:05 -0500 (0:00:00.044)       0:04:52.758 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Saturday 14 December 2024  17:59:05 -0500 (0:00:00.078)       0:04:52.836 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Saturday 14 December 2024  17:59:05 -0500 (0:00:00.058)       0:04:52.895 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217134.9187565, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "ctime": 1734217134.9187565, 
        "dev": 5, 
        "device_type": 2049, 
        "executable": false, 
        "exists": true, 
        "gid": 6, 
        "gr_name": "disk", 
        "inode": 422433, 
        "isblk": true, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": false, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/blockdevice", 
        "mode": "0660", 
        "mtime": 1734217134.9187565, 
        "nlink": 1, 
        "path": "/dev/sda1", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": null, 
        "wgrp": true, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Saturday 14 December 2024  17:59:06 -0500 (0:00:00.622)       0:04:53.517 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Saturday 14 December 2024  17:59:06 -0500 (0:00:00.154)       0:04:53.672 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Saturday 14 December 2024  17:59:06 -0500 (0:00:00.108)       0:04:53.780 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Saturday 14 December 2024  17:59:06 -0500 (0:00:00.101)       0:04:53.882 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "st_volume_type": "partition"
    }, 
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Saturday 14 December 2024  17:59:06 -0500 (0:00:00.090)       0:04:53.973 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Saturday 14 December 2024  17:59:06 -0500 (0:00:00.103)       0:04:54.076 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Saturday 14 December 2024  17:59:06 -0500 (0:00:00.112)       0:04:54.189 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217135.0367563, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "ctime": 1734217135.0367563, 
        "dev": 5, 
        "device_type": 64768, 
        "executable": false, 
        "exists": true, 
        "gid": 6, 
        "gr_name": "disk", 
        "inode": 421521, 
        "isblk": true, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": false, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/symlink", 
        "mode": "0660", 
        "mtime": 1734217135.0367563, 
        "nlink": 1, 
        "path": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": null, 
        "wgrp": true, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Saturday 14 December 2024  17:59:07 -0500 (0:00:00.471)       0:04:54.660 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}
lsrpackages: cryptsetup

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Saturday 14 December 2024  17:59:08 -0500 (0:00:00.976)       0:04:55.636 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cryptsetup", 
        "luksDump", 
        "/dev/sda1"
    ], 
    "delta": "0:00:00.026152", 
    "end": "2024-12-14 17:59:08.725456", 
    "rc": 0, 
    "start": "2024-12-14 17:59:08.699304"
}

STDOUT:

LUKS header information for /dev/sda1

Version:       	1
Cipher name:   	aes
Cipher mode:   	xts-plain64
Hash spec:     	sha256
Payload offset:	8192
MK bits:       	512
MK digest:     	17 92 c9 1e 79 04 84 9b 99 66 26 41 26 df 52 41 ff 52 d5 40 
MK salt:       	60 5b be 50 7c 58 cd 6b 10 d2 fd 5c 37 e3 d5 f5 
               	a4 09 77 44 36 22 d6 98 70 43 20 ab bd e0 ce d6 
MK iterations: 	23141
UUID:          	b6506dcf-75d2-4401-8040-0091a37b6dfd

Key Slot 0: ENABLED
	Iterations:         	370782
	Salt:               	b0 a1 12 a0 b0 71 84 77 26 42 38 cc b4 00 ab 03 
	                      	73 91 17 20 b8 e5 fb 6c 87 c6 7c 18 1f 0c aa 0b 
	Key material offset:	8
	AF stripes:            	4000
Key Slot 1: DISABLED
Key Slot 2: DISABLED
Key Slot 3: DISABLED
Key Slot 4: DISABLED
Key Slot 5: DISABLED
Key Slot 6: DISABLED
Key Slot 7: DISABLED

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Saturday 14 December 2024  17:59:08 -0500 (0:00:00.581)       0:04:56.218 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Saturday 14 December 2024  17:59:08 -0500 (0:00:00.097)       0:04:56.316 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Saturday 14 December 2024  17:59:09 -0500 (0:00:00.074)       0:04:56.390 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Saturday 14 December 2024  17:59:09 -0500 (0:00:00.078)       0:04:56.468 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Saturday 14 December 2024  17:59:09 -0500 (0:00:00.070)       0:04:56.539 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Saturday 14 December 2024  17:59:09 -0500 (0:00:00.066)       0:04:56.606 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Saturday 14 December 2024  17:59:09 -0500 (0:00:00.062)       0:04:56.668 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Saturday 14 December 2024  17:59:09 -0500 (0:00:00.079)       0:04:56.748 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [
            "luks-b6506dcf-75d2-4401-8040-0091a37b6dfd /dev/sda1 VALUE_SPECIFIED_IN_NO_LOG_PARAMETER"
        ], 
        "_storage_test_expected_crypttab_entries": "1", 
        "_storage_test_expected_crypttab_key_file": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER"
    }, 
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Saturday 14 December 2024  17:59:09 -0500 (0:00:00.075)       0:04:56.823 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Saturday 14 December 2024  17:59:09 -0500 (0:00:00.096)       0:04:56.920 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Saturday 14 December 2024  17:59:09 -0500 (0:00:00.149)       0:04:57.070 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Saturday 14 December 2024  17:59:09 -0500 (0:00:00.173)       0:04:57.243 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Saturday 14 December 2024  17:59:09 -0500 (0:00:00.108)       0:04:57.352 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null, 
        "_storage_test_expected_crypttab_entries": null, 
        "_storage_test_expected_crypttab_key_file": null
    }, 
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Saturday 14 December 2024  17:59:10 -0500 (0:00:00.075)       0:04:57.428 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Saturday 14 December 2024  17:59:10 -0500 (0:00:00.078)       0:04:57.507 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Saturday 14 December 2024  17:59:10 -0500 (0:00:00.057)       0:04:57.565 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Saturday 14 December 2024  17:59:10 -0500 (0:00:00.068)       0:04:57.633 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Saturday 14 December 2024  17:59:10 -0500 (0:00:00.110)       0:04:57.743 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Saturday 14 December 2024  17:59:10 -0500 (0:00:00.104)       0:04:57.847 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Saturday 14 December 2024  17:59:10 -0500 (0:00:00.097)       0:04:57.945 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Saturday 14 December 2024  17:59:10 -0500 (0:00:00.091)       0:04:58.036 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Saturday 14 December 2024  17:59:10 -0500 (0:00:00.073)       0:04:58.109 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Saturday 14 December 2024  17:59:10 -0500 (0:00:00.095)       0:04:58.205 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Saturday 14 December 2024  17:59:10 -0500 (0:00:00.113)       0:04:58.318 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Saturday 14 December 2024  17:59:10 -0500 (0:00:00.060)       0:04:58.379 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Saturday 14 December 2024  17:59:11 -0500 (0:00:00.059)       0:04:58.439 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Saturday 14 December 2024  17:59:11 -0500 (0:00:00.074)       0:04:58.513 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Saturday 14 December 2024  17:59:11 -0500 (0:00:00.098)       0:04:58.612 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Saturday 14 December 2024  17:59:11 -0500 (0:00:00.097)       0:04:58.710 ***** 
skipping: [managed-node2] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Saturday 14 December 2024  17:59:11 -0500 (0:00:00.090)       0:04:58.800 ***** 
skipping: [managed-node2] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Saturday 14 December 2024  17:59:11 -0500 (0:00:00.059)       0:04:58.859 ***** 
skipping: [managed-node2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Saturday 14 December 2024  17:59:11 -0500 (0:00:00.097)       0:04:58.957 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Saturday 14 December 2024  17:59:11 -0500 (0:00:00.118)       0:04:59.075 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Saturday 14 December 2024  17:59:11 -0500 (0:00:00.092)       0:04:59.168 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Saturday 14 December 2024  17:59:11 -0500 (0:00:00.089)       0:04:59.258 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Saturday 14 December 2024  17:59:11 -0500 (0:00:00.095)       0:04:59.353 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Saturday 14 December 2024  17:59:12 -0500 (0:00:00.113)       0:04:59.467 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Saturday 14 December 2024  17:59:12 -0500 (0:00:00.055)       0:04:59.522 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Saturday 14 December 2024  17:59:12 -0500 (0:00:00.056)       0:04:59.578 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Saturday 14 December 2024  17:59:12 -0500 (0:00:00.070)       0:04:59.648 ***** 
skipping: [managed-node2] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Saturday 14 December 2024  17:59:12 -0500 (0:00:00.084)       0:04:59.733 ***** 
skipping: [managed-node2] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Saturday 14 December 2024  17:59:12 -0500 (0:00:00.136)       0:04:59.869 ***** 
skipping: [managed-node2] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Saturday 14 December 2024  17:59:12 -0500 (0:00:00.083)       0:04:59.952 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Saturday 14 December 2024  17:59:12 -0500 (0:00:00.128)       0:05:00.081 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Saturday 14 December 2024  17:59:12 -0500 (0:00:00.111)       0:05:00.192 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Saturday 14 December 2024  17:59:12 -0500 (0:00:00.117)       0:05:00.309 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Saturday 14 December 2024  17:59:13 -0500 (0:00:00.081)       0:05:00.391 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Saturday 14 December 2024  17:59:13 -0500 (0:00:00.104)       0:05:00.495 ***** 
ok: [managed-node2] => {
    "storage_test_actual_size": {
        "changed": false, 
        "skip_reason": "Conditional result was False", 
        "skipped": true
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Saturday 14 December 2024  17:59:13 -0500 (0:00:00.095)       0:05:00.590 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Saturday 14 December 2024  17:59:13 -0500 (0:00:00.059)       0:05:00.650 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Saturday 14 December 2024  17:59:13 -0500 (0:00:00.110)       0:05:00.761 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Saturday 14 December 2024  17:59:13 -0500 (0:00:00.064)       0:05:00.826 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Saturday 14 December 2024  17:59:13 -0500 (0:00:00.078)       0:05:00.904 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Saturday 14 December 2024  17:59:13 -0500 (0:00:00.071)       0:05:00.976 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Saturday 14 December 2024  17:59:13 -0500 (0:00:00.065)       0:05:01.042 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Saturday 14 December 2024  17:59:13 -0500 (0:00:00.071)       0:05:01.113 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Saturday 14 December 2024  17:59:13 -0500 (0:00:00.056)       0:05:01.169 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Saturday 14 December 2024  17:59:13 -0500 (0:00:00.055)       0:05:01.225 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    }, 
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Saturday 14 December 2024  17:59:13 -0500 (0:00:00.054)       0:05:01.279 ***** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Saturday 14 December 2024  17:59:13 -0500 (0:00:00.059)       0:05:01.338 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null, 
        "storage_test_crypttab": null, 
        "storage_test_fstab": null
    }, 
    "changed": false
}

TASK [Remove the key file] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:346
Saturday 14 December 2024  17:59:14 -0500 (0:00:00.060)       0:05:01.399 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "path": "/tmp/storage_test3FfHdzlukskey", 
    "state": "absent"
}

TASK [Test for correct handling of new encrypted volume w/ no key] *************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:356
Saturday 14 December 2024  17:59:14 -0500 (0:00:00.398)       0:05:01.797 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node2

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Saturday 14 December 2024  17:59:14 -0500 (0:00:00.173)       0:05:01.971 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_pools_global": [], 
        "storage_safe_mode_global": true, 
        "storage_volumes_global": []
    }, 
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Saturday 14 December 2024  17:59:14 -0500 (0:00:00.065)       0:05:02.036 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  17:59:14 -0500 (0:00:00.095)       0:05:02.132 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  17:59:14 -0500 (0:00:00.093)       0:05:02.225 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  17:59:14 -0500 (0:00:00.065)       0:05:02.290 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  17:59:15 -0500 (0:00:00.133)       0:05:02.424 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  17:59:15 -0500 (0:00:00.057)       0:05:02.482 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  17:59:15 -0500 (0:00:00.060)       0:05:02.542 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  17:59:15 -0500 (0:00:00.072)       0:05:02.614 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  17:59:15 -0500 (0:00:00.060)       0:05:02.675 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  17:59:15 -0500 (0:00:00.134)       0:05:02.809 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  17:59:16 -0500 (0:00:01.196)       0:05:04.006 ***** 
ok: [managed-node2] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ], 
            "name": "foo", 
            "type": "lvm", 
            "volumes": [
                {
                    "encryption": true, 
                    "mount_point": "/opt/test1", 
                    "name": "test1", 
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  17:59:16 -0500 (0:00:00.047)       0:05:04.054 ***** 
ok: [managed-node2] => {
    "storage_volumes": []
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  17:59:16 -0500 (0:00:00.060)       0:05:04.114 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [
        "cryptsetup", 
        "lvm2"
    ], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  17:59:20 -0500 (0:00:03.966)       0:05:08.081 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  17:59:20 -0500 (0:00:00.071)       0:05:08.152 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  17:59:20 -0500 (0:00:00.035)       0:05:08.188 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  17:59:20 -0500 (0:00:00.045)       0:05:08.234 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  17:59:20 -0500 (0:00:00.048)       0:05:08.282 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed", 
        "7:lvm2-2.02.187-6.el7_9.5.x86_64 providing lvm2 is already installed", 
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: cryptsetup kpartx lvm2

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  17:59:21 -0500 (0:00:00.686)       0:05:08.968 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  17:59:22 -0500 (0:00:01.032)       0:05:10.000 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  17:59:22 -0500 (0:00:00.073)       0:05:10.073 ***** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  17:59:22 -0500 (0:00:00.043)       0:05:10.116 ***** 
fatal: [managed-node2]: FAILED! => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [], 
    "pools": [], 
    "volumes": []
}

MSG:

encrypted volume 'test1' missing key/password

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Saturday 14 December 2024  17:59:27 -0500 (0:00:04.273)       0:05:14.390 ***** 
fatal: [managed-node2]: FAILED! => {
    "changed": false
}

MSG:

{u'_ansible_no_log': False, u'crypts': [], u'pools': [], u'leaves': [], u'changed': False, u'actions': [], u'failed': True, u'volumes': [], u'invocation': {u'module_args': {u'packages_only': False, u'disklabel_type': None, u'diskvolume_mkfs_option_map': {u'ext4': u'-F', u'ext3': u'-F', u'ext2': u'-F'}, u'safe_mode': False, u'pools': [{u'raid_metadata_version': None, u'encryption_key_size': None, u'encryption_key': None, u'encryption_luks_version': None, u'encryption_tang_url': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_tang_thumbprint': None, u'name': u'foo', u'encryption_password': None, u'encryption': False, u'disks': [u'sda'], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [{u'raid_metadata_version': None, u'mount_device_identifier': u'uuid', u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': u'4g', u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': None, u'encryption': True, u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'type': u'lvm', u'encryption_key_size': None, u'deduplication': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'mount_passno': 0, u'mount_user': None, u'raid_spare_count': None, u'name': u'test1', u'cache_mode': None, u'raid_disks': [], u'mount_group': None, u'fs_overwrite_existing': True, u'disks': [], u'cached': False, u'thin_pool_size': None, u'thin': False, u'mount_check': 0, u'cache_size': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}], u'shared': False, u'encryption_clevis_pin': None, u'type': u'lvm', u'encryption_cipher': None, u'raid_chunk_size': None}], u'volumes': [], u'pool_defaults': {u'raid_metadata_version': None, u'encryption_cipher': None, u'encryption_key': None, u'encryption_luks_version': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_password': None, u'encryption': False, u'disks': [], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [], u'shared': False, u'type': u'lvm', u'encryption_key_size': None, u'raid_chunk_size': None}, u'volume_defaults': {u'raid_metadata_version': None, u'raid_level': None, u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': 0, u'mount_point': u'', u'compression': None, u'encryption_password': None, u'encryption': False, u'mount_device_identifier': u'uuid', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'thin_pool_name': None, u'type': u'lvm', u'encryption_key_size': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'cache_size': 0, u'raid_spare_count': None, u'cache_mode': None, u'deduplication': None, u'cached': False, u'fs_overwrite_existing': True, u'disks': [], u'thin_pool_size': None, u'thin': None, u'mount_check': 0, u'mount_passno': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}, u'use_partitions': None}}, u'mounts': [], u'packages': [], u'msg': u"encrypted volume 'test1' missing key/password"}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  17:59:27 -0500 (0:00:00.053)       0:05:14.443 ***** 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Saturday 14 December 2024  17:59:27 -0500 (0:00:00.036)       0:05:14.479 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Saturday 14 December 2024  17:59:27 -0500 (0:00:00.044)       0:05:14.524 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Saturday 14 December 2024  17:59:27 -0500 (0:00:00.053)       0:05:14.578 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Create an encrypted lvm volume w/ default fs] ****************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:374
Saturday 14 December 2024  17:59:27 -0500 (0:00:00.038)       0:05:14.616 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  17:59:27 -0500 (0:00:00.074)       0:05:14.691 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  17:59:27 -0500 (0:00:00.122)       0:05:14.814 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  17:59:27 -0500 (0:00:00.046)       0:05:14.861 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  17:59:27 -0500 (0:00:00.109)       0:05:14.970 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  17:59:27 -0500 (0:00:00.039)       0:05:15.009 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  17:59:27 -0500 (0:00:00.039)       0:05:15.049 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  17:59:27 -0500 (0:00:00.038)       0:05:15.088 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  17:59:27 -0500 (0:00:00.037)       0:05:15.126 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  17:59:27 -0500 (0:00:00.088)       0:05:15.215 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  17:59:29 -0500 (0:00:01.190)       0:05:16.405 ***** 
ok: [managed-node2] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ], 
            "name": "foo", 
            "type": "lvm", 
            "volumes": [
                {
                    "encryption": true, 
                    "encryption_cipher": "aes-xts-plain64", 
                    "encryption_key_size": 512, 
                    "encryption_luks_version": "luks1", 
                    "encryption_password": "yabbadabbadoo", 
                    "mount_point": "/opt/test1", 
                    "name": "test1", 
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  17:59:29 -0500 (0:00:00.053)       0:05:16.458 ***** 
ok: [managed-node2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  17:59:29 -0500 (0:00:00.046)       0:05:16.505 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [
        "cryptsetup", 
        "lvm2"
    ], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  17:59:33 -0500 (0:00:04.168)       0:05:20.673 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  17:59:33 -0500 (0:00:00.066)       0:05:20.739 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  17:59:33 -0500 (0:00:00.035)       0:05:20.775 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  17:59:33 -0500 (0:00:00.037)       0:05:20.813 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  17:59:33 -0500 (0:00:00.034)       0:05:20.847 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed", 
        "7:lvm2-2.02.187-6.el7_9.5.x86_64 providing lvm2 is already installed", 
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: cryptsetup kpartx lvm2

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  17:59:34 -0500 (0:00:00.631)       0:05:21.479 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  17:59:35 -0500 (0:00:00.968)       0:05:22.447 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  17:59:35 -0500 (0:00:00.056)       0:05:22.504 ***** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  17:59:35 -0500 (0:00:00.041)       0:05:22.545 ***** 
changed: [managed-node2] => {
    "actions": [
        {
            "action": "destroy format", 
            "device": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
            "fs_type": "xfs"
        }, 
        {
            "action": "destroy device", 
            "device": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
            "fs_type": null
        }, 
        {
            "action": "destroy format", 
            "device": "/dev/sda1", 
            "fs_type": "luks"
        }, 
        {
            "action": "destroy device", 
            "device": "/dev/sda1", 
            "fs_type": null
        }, 
        {
            "action": "destroy format", 
            "device": "/dev/sda", 
            "fs_type": "disklabel"
        }, 
        {
            "action": "create format", 
            "device": "/dev/sda", 
            "fs_type": "lvmpv"
        }, 
        {
            "action": "create device", 
            "device": "/dev/foo", 
            "fs_type": null
        }, 
        {
            "action": "create device", 
            "device": "/dev/mapper/foo-test1", 
            "fs_type": null
        }, 
        {
            "action": "create format", 
            "device": "/dev/mapper/foo-test1", 
            "fs_type": "luks"
        }, 
        {
            "action": "create device", 
            "device": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
            "fs_type": null
        }, 
        {
            "action": "create format", 
            "device": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
            "fs_type": "xfs"
        }
    ], 
    "changed": true, 
    "crypts": [
        {
            "backing_device": "/dev/sda1", 
            "name": "luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
            "password": "-", 
            "state": "absent"
        }, 
        {
            "backing_device": "/dev/mapper/foo-test1", 
            "name": "luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
            "password": "-", 
            "state": "present"
        }
    ], 
    "leaves": [
        "/dev/sdb", 
        "/dev/sdc", 
        "/dev/sdd", 
        "/dev/sde", 
        "/dev/sdf", 
        "/dev/sdg", 
        "/dev/sdh", 
        "/dev/sdi", 
        "/dev/xvda1", 
        "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c"
    ], 
    "mounts": [
        {
            "fstype": "xfs", 
            "path": "/opt/test1", 
            "src": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
            "state": "absent"
        }, 
        {
            "dump": 0, 
            "fstype": "xfs", 
            "group": null, 
            "mode": null, 
            "opts": "defaults", 
            "owner": null, 
            "passno": 0, 
            "path": "/opt/test1", 
            "src": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
            "state": "mounted"
        }
    ], 
    "packages": [
        "cryptsetup", 
        "xfsprogs", 
        "e2fsprogs", 
        "lvm2"
    ], 
    "pools": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_cipher": null, 
            "encryption_clevis_pin": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": null, 
            "encryption_tang_thumbprint": null, 
            "encryption_tang_url": null, 
            "grow_to_fill": false, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "shared": false, 
            "state": "present", 
            "type": "lvm", 
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                    "_kernel_device": "/dev/dm-1", 
                    "_mount_id": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                    "_raw_device": "/dev/mapper/foo-test1", 
                    "_raw_kernel_device": "/dev/dm-0", 
                    "cache_devices": [], 
                    "cache_mode": null, 
                    "cache_size": 0, 
                    "cached": false, 
                    "compression": null, 
                    "deduplication": null, 
                    "disks": [], 
                    "encryption": true, 
                    "encryption_cipher": "aes-xts-plain64", 
                    "encryption_key": null, 
                    "encryption_key_size": 512, 
                    "encryption_luks_version": "luks1", 
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                    "fs_create_options": "", 
                    "fs_label": "", 
                    "fs_overwrite_existing": true, 
                    "fs_type": "xfs", 
                    "mount_check": 0, 
                    "mount_device_identifier": "uuid", 
                    "mount_group": null, 
                    "mount_mode": null, 
                    "mount_options": "defaults", 
                    "mount_passno": 0, 
                    "mount_point": "/opt/test1", 
                    "mount_user": null, 
                    "name": "test1", 
                    "raid_chunk_size": null, 
                    "raid_device_count": null, 
                    "raid_disks": [], 
                    "raid_level": null, 
                    "raid_metadata_version": null, 
                    "raid_spare_count": null, 
                    "raid_stripe_size": null, 
                    "size": "4g", 
                    "state": "present", 
                    "thin": false, 
                    "thin_pool_name": null, 
                    "thin_pool_size": null, 
                    "type": "lvm", 
                    "vdo_pool_size": null
                }
            ]
        }
    ], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Saturday 14 December 2024  17:59:45 -0500 (0:00:10.767)       0:05:33.312 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Saturday 14 December 2024  17:59:45 -0500 (0:00:00.046)       0:05:33.359 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217137.6327553, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "31da118dcc3004e61c657402f6c0dd12a730c28a", 
        "ctime": 1734217137.628755, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 263588, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0644", 
        "mtime": 1734217137.628755, 
        "nlink": 1, 
        "path": "/etc/fstab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 1311, 
        "uid": 0, 
        "version": "606402765", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Saturday 14 December 2024  17:59:46 -0500 (0:00:00.340)       0:05:33.699 ***** 
ok: [managed-node2] => {
    "backup": "", 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  17:59:46 -0500 (0:00:00.335)       0:05:34.034 ***** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Saturday 14 December 2024  17:59:46 -0500 (0:00:00.034)       0:05:34.068 ***** 
ok: [managed-node2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format", 
                "device": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
                "fs_type": "xfs"
            }, 
            {
                "action": "destroy device", 
                "device": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
                "fs_type": null
            }, 
            {
                "action": "destroy format", 
                "device": "/dev/sda1", 
                "fs_type": "luks"
            }, 
            {
                "action": "destroy device", 
                "device": "/dev/sda1", 
                "fs_type": null
            }, 
            {
                "action": "destroy format", 
                "device": "/dev/sda", 
                "fs_type": "disklabel"
            }, 
            {
                "action": "create format", 
                "device": "/dev/sda", 
                "fs_type": "lvmpv"
            }, 
            {
                "action": "create device", 
                "device": "/dev/foo", 
                "fs_type": null
            }, 
            {
                "action": "create device", 
                "device": "/dev/mapper/foo-test1", 
                "fs_type": null
            }, 
            {
                "action": "create format", 
                "device": "/dev/mapper/foo-test1", 
                "fs_type": "luks"
            }, 
            {
                "action": "create device", 
                "device": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                "fs_type": null
            }, 
            {
                "action": "create format", 
                "device": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                "fs_type": "xfs"
            }
        ], 
        "changed": true, 
        "crypts": [
            {
                "backing_device": "/dev/sda1", 
                "name": "luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
                "password": "-", 
                "state": "absent"
            }, 
            {
                "backing_device": "/dev/mapper/foo-test1", 
                "name": "luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                "password": "-", 
                "state": "present"
            }
        ], 
        "failed": false, 
        "leaves": [
            "/dev/sdb", 
            "/dev/sdc", 
            "/dev/sdd", 
            "/dev/sde", 
            "/dev/sdf", 
            "/dev/sdg", 
            "/dev/sdh", 
            "/dev/sdi", 
            "/dev/xvda1", 
            "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c"
        ], 
        "mounts": [
            {
                "fstype": "xfs", 
                "path": "/opt/test1", 
                "src": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
                "state": "absent"
            }, 
            {
                "dump": 0, 
                "fstype": "xfs", 
                "group": null, 
                "mode": null, 
                "opts": "defaults", 
                "owner": null, 
                "passno": 0, 
                "path": "/opt/test1", 
                "src": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                "state": "mounted"
            }
        ], 
        "packages": [
            "cryptsetup", 
            "xfsprogs", 
            "e2fsprogs", 
            "lvm2"
        ], 
        "pools": [
            {
                "disks": [
                    "sda"
                ], 
                "encryption": false, 
                "encryption_cipher": null, 
                "encryption_clevis_pin": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": null, 
                "encryption_tang_thumbprint": null, 
                "encryption_tang_url": null, 
                "grow_to_fill": false, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "shared": false, 
                "state": "present", 
                "type": "lvm", 
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                        "_kernel_device": "/dev/dm-1", 
                        "_mount_id": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                        "_raw_device": "/dev/mapper/foo-test1", 
                        "_raw_kernel_device": "/dev/dm-0", 
                        "cache_devices": [], 
                        "cache_mode": null, 
                        "cache_size": 0, 
                        "cached": false, 
                        "compression": null, 
                        "deduplication": null, 
                        "disks": [], 
                        "encryption": true, 
                        "encryption_cipher": "aes-xts-plain64", 
                        "encryption_key": null, 
                        "encryption_key_size": 512, 
                        "encryption_luks_version": "luks1", 
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                        "fs_create_options": "", 
                        "fs_label": "", 
                        "fs_overwrite_existing": true, 
                        "fs_type": "xfs", 
                        "mount_check": 0, 
                        "mount_device_identifier": "uuid", 
                        "mount_group": null, 
                        "mount_mode": null, 
                        "mount_options": "defaults", 
                        "mount_passno": 0, 
                        "mount_point": "/opt/test1", 
                        "mount_user": null, 
                        "name": "test1", 
                        "raid_chunk_size": null, 
                        "raid_device_count": null, 
                        "raid_disks": [], 
                        "raid_level": null, 
                        "raid_metadata_version": null, 
                        "raid_spare_count": null, 
                        "raid_stripe_size": null, 
                        "size": "4g", 
                        "state": "present", 
                        "thin": false, 
                        "thin_pool_name": null, 
                        "thin_pool_size": null, 
                        "type": "lvm", 
                        "vdo_pool_size": null
                    }
                ]
            }
        ], 
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Saturday 14 December 2024  17:59:46 -0500 (0:00:00.060)       0:05:34.129 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ], 
                "encryption": false, 
                "encryption_cipher": null, 
                "encryption_clevis_pin": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": null, 
                "encryption_tang_thumbprint": null, 
                "encryption_tang_url": null, 
                "grow_to_fill": false, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "shared": false, 
                "state": "present", 
                "type": "lvm", 
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                        "_kernel_device": "/dev/dm-1", 
                        "_mount_id": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                        "_raw_device": "/dev/mapper/foo-test1", 
                        "_raw_kernel_device": "/dev/dm-0", 
                        "cache_devices": [], 
                        "cache_mode": null, 
                        "cache_size": 0, 
                        "cached": false, 
                        "compression": null, 
                        "deduplication": null, 
                        "disks": [], 
                        "encryption": true, 
                        "encryption_cipher": "aes-xts-plain64", 
                        "encryption_key": null, 
                        "encryption_key_size": 512, 
                        "encryption_luks_version": "luks1", 
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                        "fs_create_options": "", 
                        "fs_label": "", 
                        "fs_overwrite_existing": true, 
                        "fs_type": "xfs", 
                        "mount_check": 0, 
                        "mount_device_identifier": "uuid", 
                        "mount_group": null, 
                        "mount_mode": null, 
                        "mount_options": "defaults", 
                        "mount_passno": 0, 
                        "mount_point": "/opt/test1", 
                        "mount_user": null, 
                        "name": "test1", 
                        "raid_chunk_size": null, 
                        "raid_device_count": null, 
                        "raid_disks": [], 
                        "raid_level": null, 
                        "raid_metadata_version": null, 
                        "raid_spare_count": null, 
                        "raid_stripe_size": null, 
                        "size": "4g", 
                        "state": "present", 
                        "thin": false, 
                        "thin_pool_name": null, 
                        "thin_pool_size": null, 
                        "type": "lvm", 
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Saturday 14 December 2024  17:59:46 -0500 (0:00:00.063)       0:05:34.193 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Saturday 14 December 2024  17:59:46 -0500 (0:00:00.068)       0:05:34.262 ***** 
changed: [managed-node2] => (item={u'src': u'/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd', u'state': u'absent', u'fstype': u'xfs', u'path': u'/opt/test1'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": true, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "fstype": "xfs", 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
        "state": "absent"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "/dev/mapper/luks-b6506dcf-75d2-4401-8040-0091a37b6dfd"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Saturday 14 December 2024  17:59:47 -0500 (0:00:00.407)       0:05:34.669 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Saturday 14 December 2024  17:59:47 -0500 (0:00:00.513)       0:05:35.183 ***** 
changed: [managed-node2] => (item={u'src': u'/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": true, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
        "state": "mounted"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Saturday 14 December 2024  17:59:48 -0500 (0:00:00.407)       0:05:35.591 ***** 
skipping: [managed-node2] => (item={u'src': u'/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'})  => {
    "ansible_loop_var": "mount_info", 
    "changed": false, 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
        "state": "mounted"
    }, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Saturday 14 December 2024  17:59:48 -0500 (0:00:00.108)       0:05:35.699 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Saturday 14 December 2024  17:59:48 -0500 (0:00:00.490)       0:05:36.190 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217140.9657536, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "ed919d0d695d56489e27497f0e47d9445d78699b", 
        "ctime": 1734217138.9907546, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 917512, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0600", 
        "mtime": 1734217138.9907546, 
        "nlink": 1, 
        "path": "/etc/crypttab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": false, 
        "roth": false, 
        "rusr": true, 
        "size": 88, 
        "uid": 0, 
        "version": "606409399", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Saturday 14 December 2024  17:59:49 -0500 (0:00:00.332)       0:05:36.523 ***** 
changed: [managed-node2] => (item={u'state': u'absent', u'password': u'-', u'name': u'luks-b6506dcf-75d2-4401-8040-0091a37b6dfd', u'backing_device': u'/dev/sda1'}) => {
    "ansible_loop_var": "entry", 
    "backup": "", 
    "changed": true, 
    "entry": {
        "backing_device": "/dev/sda1", 
        "name": "luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
        "password": "-", 
        "state": "absent"
    }, 
    "found": 1
}

MSG:

1 line(s) removed
changed: [managed-node2] => (item={u'state': u'present', u'password': u'-', u'name': u'luks-3365bc93-2fc0-47e8-b784-b63a2126f78c', u'backing_device': u'/dev/mapper/foo-test1'}) => {
    "ansible_loop_var": "entry", 
    "backup": "", 
    "changed": true, 
    "entry": {
        "backing_device": "/dev/mapper/foo-test1", 
        "name": "luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
        "password": "-", 
        "state": "present"
    }
}

MSG:

line added

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Saturday 14 December 2024  17:59:49 -0500 (0:00:00.708)       0:05:37.232 ***** 
ok: [managed-node2]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:393
Saturday 14 December 2024  17:59:50 -0500 (0:00:00.718)       0:05:37.950 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node2

TASK [Print out pool information] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Saturday 14 December 2024  17:59:50 -0500 (0:00:00.105)       0:05:38.055 ***** 
ok: [managed-node2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_cipher": null, 
            "encryption_clevis_pin": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": null, 
            "encryption_tang_thumbprint": null, 
            "encryption_tang_url": null, 
            "grow_to_fill": false, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "shared": false, 
            "state": "present", 
            "type": "lvm", 
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                    "_kernel_device": "/dev/dm-1", 
                    "_mount_id": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                    "_raw_device": "/dev/mapper/foo-test1", 
                    "_raw_kernel_device": "/dev/dm-0", 
                    "cache_devices": [], 
                    "cache_mode": null, 
                    "cache_size": 0, 
                    "cached": false, 
                    "compression": null, 
                    "deduplication": null, 
                    "disks": [], 
                    "encryption": true, 
                    "encryption_cipher": "aes-xts-plain64", 
                    "encryption_key": null, 
                    "encryption_key_size": 512, 
                    "encryption_luks_version": "luks1", 
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                    "fs_create_options": "", 
                    "fs_label": "", 
                    "fs_overwrite_existing": true, 
                    "fs_type": "xfs", 
                    "mount_check": 0, 
                    "mount_device_identifier": "uuid", 
                    "mount_group": null, 
                    "mount_mode": null, 
                    "mount_options": "defaults", 
                    "mount_passno": 0, 
                    "mount_point": "/opt/test1", 
                    "mount_user": null, 
                    "name": "test1", 
                    "raid_chunk_size": null, 
                    "raid_device_count": null, 
                    "raid_disks": [], 
                    "raid_level": null, 
                    "raid_metadata_version": null, 
                    "raid_spare_count": null, 
                    "raid_stripe_size": null, 
                    "size": "4g", 
                    "state": "present", 
                    "thin": false, 
                    "thin_pool_name": null, 
                    "thin_pool_size": null, 
                    "type": "lvm", 
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Saturday 14 December 2024  17:59:50 -0500 (0:00:00.075)       0:05:38.130 ***** 
skipping: [managed-node2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Saturday 14 December 2024  17:59:50 -0500 (0:00:00.054)       0:05:38.185 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "crypto_LUKS", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/mapper/foo-test1", 
            "size": "4G", 
            "type": "lvm", 
            "uuid": "3365bc93-2fc0-47e8-b784-b63a2126f78c"
        }, 
        "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c": {
            "fstype": "xfs", 
            "label": "", 
            "mountpoint": "/opt/test1", 
            "name": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
            "size": "4G", 
            "type": "crypt", 
            "uuid": "4f26c9ff-d55a-4463-b590-1cd68a73451b"
        }, 
        "/dev/sda": {
            "fstype": "LVM2_member", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sda", 
            "size": "10G", 
            "type": "disk", 
            "uuid": "xa7q16-X4RQ-vFWY-Ftcj-AFkA-0I7B-z0LDg5"
        }, 
        "/dev/sdb": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdb", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdc": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdc", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdd": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdd", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sde": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sde", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdf": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdf", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdg": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdg", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdh": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdh", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdi": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdi", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/xvda", 
            "size": "250G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda1": {
            "fstype": "ext4", 
            "label": "", 
            "mountpoint": "/", 
            "name": "/dev/xvda1", 
            "size": "250G", 
            "type": "partition", 
            "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Saturday 14 December 2024  17:59:51 -0500 (0:00:00.389)       0:05:38.575 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/fstab"
    ], 
    "delta": "0:00:00.002953", 
    "end": "2024-12-14 17:59:51.470260", 
    "rc": 0, 
    "start": "2024-12-14 17:59:51.467307"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Thu Jun 20 10:23:46 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 /                       ext4    defaults        1 1
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Saturday 14 December 2024  17:59:51 -0500 (0:00:00.354)       0:05:38.929 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/crypttab"
    ], 
    "delta": "0:00:00.002865", 
    "end": "2024-12-14 17:59:51.816711", 
    "failed_when_result": false, 
    "rc": 0, 
    "start": "2024-12-14 17:59:51.813846"
}

STDOUT:

luks-3365bc93-2fc0-47e8-b784-b63a2126f78c /dev/mapper/foo-test1 -

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Saturday 14 December 2024  17:59:51 -0500 (0:00:00.332)       0:05:39.262 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node2

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Saturday 14 December 2024  17:59:51 -0500 (0:00:00.083)       0:05:39.345 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members", 
            "volumes"
        ]
    }, 
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Saturday 14 December 2024  17:59:52 -0500 (0:00:00.038)       0:05:39.384 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "vgs", 
        "--noheadings", 
        "--binary", 
        "-o", 
        "shared", 
        "foo"
    ], 
    "delta": "0:00:00.021642", 
    "end": "2024-12-14 17:59:52.315637", 
    "rc": 0, 
    "start": "2024-12-14 17:59:52.293995"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Saturday 14 December 2024  17:59:52 -0500 (0:00:00.401)       0:05:39.785 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Saturday 14 December 2024  17:59:52 -0500 (0:00:00.061)       0:05:39.847 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node2

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Saturday 14 December 2024  17:59:52 -0500 (0:00:00.107)       0:05:39.954 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1", 
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    }, 
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Saturday 14 December 2024  17:59:52 -0500 (0:00:00.050)       0:05:40.005 ***** 
ok: [managed-node2] => (item=/dev/sda) => {
    "ansible_loop_var": "pv", 
    "changed": false, 
    "device": "/dev/sda", 
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Saturday 14 December 2024  17:59:53 -0500 (0:00:00.550)       0:05:40.555 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    }, 
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Saturday 14 December 2024  17:59:53 -0500 (0:00:00.056)       0:05:40.612 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    }, 
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Saturday 14 December 2024  17:59:53 -0500 (0:00:00.056)       0:05:40.668 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Saturday 14 December 2024  17:59:53 -0500 (0:00:00.060)       0:05:40.729 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    }, 
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Saturday 14 December 2024  17:59:53 -0500 (0:00:00.043)       0:05:40.773 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    }, 
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Saturday 14 December 2024  17:59:53 -0500 (0:00:00.047)       0:05:40.820 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Saturday 14 December 2024  17:59:53 -0500 (0:00:00.052)       0:05:40.873 ***** 
ok: [managed-node2] => (item=/dev/sda) => {
    "ansible_loop_var": "pv", 
    "changed": false, 
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Saturday 14 December 2024  17:59:53 -0500 (0:00:00.072)       0:05:40.946 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.8.171 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Saturday 14 December 2024  17:59:53 -0500 (0:00:00.304)       0:05:41.250 ***** 
skipping: [managed-node2] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv", 
    "changed": false, 
    "skip_reason": "Conditional result was False", 
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Saturday 14 December 2024  17:59:53 -0500 (0:00:00.048)       0:05:41.299 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node2

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.095)       0:05:41.394 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.040)       0:05:41.435 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.037)       0:05:41.472 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.036)       0:05:41.509 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.039)       0:05:41.549 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.097)       0:05:41.646 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.038)       0:05:41.684 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.039)       0:05:41.724 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.048)       0:05:41.772 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.037)       0:05:41.810 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.037)       0:05:41.848 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null, 
        "storage_test_md_chunk_size_re": null, 
        "storage_test_md_metadata_version_re": null, 
        "storage_test_md_spare_devices_re": null
    }, 
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.040)       0:05:41.888 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.077)       0:05:41.965 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node2

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.085)       0:05:42.051 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.038)       0:05:42.089 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.036)       0:05:42.126 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.037)       0:05:42.164 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.037)       0:05:42.201 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.037)       0:05:42.239 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.052)       0:05:42.291 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Saturday 14 December 2024  17:59:54 -0500 (0:00:00.057)       0:05:42.349 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Saturday 14 December 2024  17:59:55 -0500 (0:00:00.124)       0:05:42.473 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node2

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Saturday 14 December 2024  17:59:55 -0500 (0:00:00.121)       0:05:42.595 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Saturday 14 December 2024  17:59:55 -0500 (0:00:00.056)       0:05:42.651 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Saturday 14 December 2024  17:59:55 -0500 (0:00:00.055)       0:05:42.707 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Saturday 14 December 2024  17:59:55 -0500 (0:00:00.058)       0:05:42.766 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    }, 
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Saturday 14 December 2024  17:59:55 -0500 (0:00:00.060)       0:05:42.827 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node2

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Saturday 14 December 2024  17:59:55 -0500 (0:00:00.139)       0:05:42.966 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0", 
        "_storage_test_expected_crypttab_key_file": "-"
    }, 
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Saturday 14 December 2024  17:59:55 -0500 (0:00:00.069)       0:05:43.035 ***** 
skipping: [managed-node2] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda", 
    "ansible_loop_var": "_storage_test_pool_member_path", 
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Saturday 14 December 2024  17:59:55 -0500 (0:00:00.068)       0:05:43.103 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node2

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Saturday 14 December 2024  17:59:55 -0500 (0:00:00.125)       0:05:43.229 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    }, 
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Saturday 14 December 2024  17:59:55 -0500 (0:00:00.083)       0:05:43.313 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Saturday 14 December 2024  17:59:56 -0500 (0:00:00.104)       0:05:43.417 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Saturday 14 December 2024  17:59:56 -0500 (0:00:00.111)       0:05:43.528 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Saturday 14 December 2024  17:59:56 -0500 (0:00:00.055)       0:05:43.584 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Saturday 14 December 2024  17:59:56 -0500 (0:00:00.071)       0:05:43.656 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    }, 
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Saturday 14 December 2024  17:59:56 -0500 (0:00:00.067)       0:05:43.723 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null, 
        "_storage_test_crypttab_key_file": null
    }, 
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Saturday 14 December 2024  17:59:56 -0500 (0:00:00.080)       0:05:43.804 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Saturday 14 December 2024  17:59:56 -0500 (0:00:00.151)       0:05:43.955 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node2

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Saturday 14 December 2024  17:59:56 -0500 (0:00:00.153)       0:05:44.109 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Saturday 14 December 2024  17:59:56 -0500 (0:00:00.059)       0:05:44.168 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Saturday 14 December 2024  17:59:56 -0500 (0:00:00.120)       0:05:44.289 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Saturday 14 December 2024  17:59:56 -0500 (0:00:00.075)       0:05:44.364 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Saturday 14 December 2024  17:59:57 -0500 (0:00:00.056)       0:05:44.421 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Saturday 14 December 2024  17:59:57 -0500 (0:00:00.060)       0:05:44.482 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Saturday 14 December 2024  17:59:57 -0500 (0:00:00.058)       0:05:44.540 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    }, 
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Saturday 14 December 2024  17:59:57 -0500 (0:00:00.047)       0:05:44.588 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node2

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Saturday 14 December 2024  17:59:57 -0500 (0:00:00.108)       0:05:44.696 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Saturday 14 December 2024  17:59:57 -0500 (0:00:00.055)       0:05:44.752 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Saturday 14 December 2024  17:59:57 -0500 (0:00:00.061)       0:05:44.813 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Saturday 14 December 2024  17:59:57 -0500 (0:00:00.079)       0:05:44.893 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Saturday 14 December 2024  17:59:57 -0500 (0:00:00.055)       0:05:44.949 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Saturday 14 December 2024  17:59:57 -0500 (0:00:00.051)       0:05:45.000 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    }, 
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Saturday 14 December 2024  17:59:57 -0500 (0:00:00.062)       0:05:45.063 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null, 
        "_storage_test_expected_pv_count": null, 
        "_storage_test_expected_pv_type": null, 
        "_storage_test_pool_pvs": [], 
        "_storage_test_pool_pvs_lvm": []
    }, 
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Saturday 14 December 2024  17:59:57 -0500 (0:00:00.154)       0:05:45.218 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node2

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Saturday 14 December 2024  17:59:57 -0500 (0:00:00.133)       0:05:45.351 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true, 
        "_storage_volume_tests": [
            "mount", 
            "fstab", 
            "fs", 
            "device", 
            "encryption", 
            "md", 
            "size", 
            "cache"
        ]
    }, 
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Saturday 14 December 2024  17:59:58 -0500 (0:00:00.071)       0:05:45.423 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node2

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Saturday 14 December 2024  17:59:58 -0500 (0:00:00.259)       0:05:45.682 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c"
    }, 
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Saturday 14 December 2024  17:59:58 -0500 (0:00:00.057)       0:05:45.740 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1", 
        "storage_test_swap_expected_matches": "0"
    }, 
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Saturday 14 December 2024  17:59:58 -0500 (0:00:00.057)       0:05:45.797 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Saturday 14 December 2024  17:59:58 -0500 (0:00:00.043)       0:05:45.841 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Saturday 14 December 2024  17:59:58 -0500 (0:00:00.057)       0:05:45.899 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Saturday 14 December 2024  17:59:58 -0500 (0:00:00.038)       0:05:45.937 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Saturday 14 December 2024  17:59:58 -0500 (0:00:00.048)       0:05:45.986 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Saturday 14 December 2024  17:59:58 -0500 (0:00:00.056)       0:05:46.042 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Saturday 14 December 2024  17:59:58 -0500 (0:00:00.068)       0:05:46.111 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Saturday 14 December 2024  17:59:58 -0500 (0:00:00.129)       0:05:46.240 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Saturday 14 December 2024  17:59:58 -0500 (0:00:00.114)       0:05:46.355 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null, 
        "storage_test_mount_expected_mount_point": null, 
        "storage_test_swap_expected_matches": null, 
        "storage_test_swaps": null, 
        "storage_test_sys_node": null
    }, 
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Saturday 14 December 2024  17:59:59 -0500 (0:00:00.100)       0:05:46.455 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1", 
        "storage_test_fstab_expected_mount_options_matches": "1", 
        "storage_test_fstab_expected_mount_point_matches": "1", 
        "storage_test_fstab_id_matches": [
            "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c "
        ], 
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ], 
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    }, 
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Saturday 14 December 2024  17:59:59 -0500 (0:00:00.122)       0:05:46.578 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Saturday 14 December 2024  17:59:59 -0500 (0:00:00.085)       0:05:46.663 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Saturday 14 December 2024  17:59:59 -0500 (0:00:00.064)       0:05:46.727 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Saturday 14 December 2024  17:59:59 -0500 (0:00:00.058)       0:05:46.786 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Saturday 14 December 2024  17:59:59 -0500 (0:00:00.069)       0:05:46.855 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null, 
        "storage_test_fstab_expected_mount_options_matches": null, 
        "storage_test_fstab_expected_mount_point_matches": null, 
        "storage_test_fstab_id_matches": null, 
        "storage_test_fstab_mount_options_matches": null, 
        "storage_test_fstab_mount_point_matches": null
    }, 
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Saturday 14 December 2024  17:59:59 -0500 (0:00:00.067)       0:05:46.923 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Saturday 14 December 2024  17:59:59 -0500 (0:00:00.101)       0:05:47.024 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Saturday 14 December 2024  17:59:59 -0500 (0:00:00.067)       0:05:47.092 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217185.683733, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "ctime": 1734217185.683733, 
        "dev": 5, 
        "device_type": 64768, 
        "executable": false, 
        "exists": true, 
        "gid": 6, 
        "gr_name": "disk", 
        "inode": 431513, 
        "isblk": true, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": false, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/symlink", 
        "mode": "0660", 
        "mtime": 1734217185.683733, 
        "nlink": 1, 
        "path": "/dev/mapper/foo-test1", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": null, 
        "wgrp": true, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Saturday 14 December 2024  18:00:00 -0500 (0:00:00.454)       0:05:47.547 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Saturday 14 December 2024  18:00:00 -0500 (0:00:00.104)       0:05:47.652 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Saturday 14 December 2024  18:00:00 -0500 (0:00:00.078)       0:05:47.730 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Saturday 14 December 2024  18:00:00 -0500 (0:00:00.087)       0:05:47.818 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    }, 
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Saturday 14 December 2024  18:00:00 -0500 (0:00:00.062)       0:05:47.881 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Saturday 14 December 2024  18:00:00 -0500 (0:00:00.057)       0:05:47.939 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Saturday 14 December 2024  18:00:00 -0500 (0:00:00.069)       0:05:48.008 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217185.796733, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "ctime": 1734217185.796733, 
        "dev": 5, 
        "device_type": 64769, 
        "executable": false, 
        "exists": true, 
        "gid": 6, 
        "gr_name": "disk", 
        "inode": 431588, 
        "isblk": true, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": false, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/symlink", 
        "mode": "0660", 
        "mtime": 1734217185.796733, 
        "nlink": 1, 
        "path": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": null, 
        "wgrp": true, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Saturday 14 December 2024  18:00:01 -0500 (0:00:00.409)       0:05:48.417 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}
lsrpackages: cryptsetup

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Saturday 14 December 2024  18:00:01 -0500 (0:00:00.778)       0:05:49.196 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cryptsetup", 
        "luksDump", 
        "/dev/mapper/foo-test1"
    ], 
    "delta": "0:00:00.026938", 
    "end": "2024-12-14 18:00:02.218270", 
    "rc": 0, 
    "start": "2024-12-14 18:00:02.191332"
}

STDOUT:

LUKS header information for /dev/mapper/foo-test1

Version:       	1
Cipher name:   	aes
Cipher mode:   	xts-plain64
Hash spec:     	sha256
Payload offset:	8192
MK bits:       	512
MK digest:     	6a ba c5 ed 79 fa 96 8d 6d d0 43 0a e8 5a 19 5c 6b 61 81 cd 
MK salt:       	40 98 5d e4 16 55 f3 9b 15 76 b2 1a b3 93 d7 4b 
               	f7 e0 a6 ec c0 1a ec 43 d0 ab 5a a5 41 ea b8 8b 
MK iterations: 	23272
UUID:          	3365bc93-2fc0-47e8-b784-b63a2126f78c

Key Slot 0: ENABLED
	Iterations:         	372362
	Salt:               	32 49 4d 0e a0 ac 4e be e7 c0 da 34 86 ff b4 98 
	                      	d4 e7 55 37 99 03 47 7d f9 b2 fb 14 94 4a a7 f1 
	Key material offset:	8
	AF stripes:            	4000
Key Slot 1: DISABLED
Key Slot 2: DISABLED
Key Slot 3: DISABLED
Key Slot 4: DISABLED
Key Slot 5: DISABLED
Key Slot 6: DISABLED
Key Slot 7: DISABLED

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Saturday 14 December 2024  18:00:02 -0500 (0:00:00.479)       0:05:49.675 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Saturday 14 December 2024  18:00:02 -0500 (0:00:00.060)       0:05:49.736 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Saturday 14 December 2024  18:00:02 -0500 (0:00:00.071)       0:05:49.807 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Saturday 14 December 2024  18:00:02 -0500 (0:00:00.067)       0:05:49.875 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Saturday 14 December 2024  18:00:02 -0500 (0:00:00.066)       0:05:49.942 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Saturday 14 December 2024  18:00:02 -0500 (0:00:00.122)       0:05:50.064 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Saturday 14 December 2024  18:00:02 -0500 (0:00:00.121)       0:05:50.186 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Saturday 14 December 2024  18:00:02 -0500 (0:00:00.158)       0:05:50.345 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [
            "luks-3365bc93-2fc0-47e8-b784-b63a2126f78c /dev/mapper/foo-test1 -"
        ], 
        "_storage_test_expected_crypttab_entries": "1", 
        "_storage_test_expected_crypttab_key_file": "-"
    }, 
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Saturday 14 December 2024  18:00:03 -0500 (0:00:00.131)       0:05:50.476 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Saturday 14 December 2024  18:00:03 -0500 (0:00:00.147)       0:05:50.624 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Saturday 14 December 2024  18:00:03 -0500 (0:00:00.099)       0:05:50.724 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Saturday 14 December 2024  18:00:03 -0500 (0:00:00.083)       0:05:50.807 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Saturday 14 December 2024  18:00:03 -0500 (0:00:00.089)       0:05:50.897 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null, 
        "_storage_test_expected_crypttab_entries": null, 
        "_storage_test_expected_crypttab_key_file": null
    }, 
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Saturday 14 December 2024  18:00:03 -0500 (0:00:00.059)       0:05:50.957 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Saturday 14 December 2024  18:00:03 -0500 (0:00:00.083)       0:05:51.040 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Saturday 14 December 2024  18:00:03 -0500 (0:00:00.056)       0:05:51.097 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Saturday 14 December 2024  18:00:03 -0500 (0:00:00.056)       0:05:51.153 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Saturday 14 December 2024  18:00:03 -0500 (0:00:00.077)       0:05:51.231 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Saturday 14 December 2024  18:00:03 -0500 (0:00:00.067)       0:05:51.299 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Saturday 14 December 2024  18:00:03 -0500 (0:00:00.064)       0:05:51.363 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Saturday 14 December 2024  18:00:04 -0500 (0:00:00.068)       0:05:51.432 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Saturday 14 December 2024  18:00:04 -0500 (0:00:00.076)       0:05:51.509 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Saturday 14 December 2024  18:00:04 -0500 (0:00:00.061)       0:05:51.571 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Saturday 14 December 2024  18:00:04 -0500 (0:00:00.055)       0:05:51.626 ***** 
ok: [managed-node2] => {
    "bytes": 4294967296, 
    "changed": false, 
    "lvm": "4g", 
    "parted": "4GiB", 
    "size": "4 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Saturday 14 December 2024  18:00:04 -0500 (0:00:00.714)       0:05:52.340 ***** 
ok: [managed-node2] => {
    "bytes": 4294967296, 
    "changed": false, 
    "lvm": "4g", 
    "parted": "4GiB", 
    "size": "4 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Saturday 14 December 2024  18:00:05 -0500 (0:00:00.488)       0:05:52.829 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_expected_size": "4294967296"
    }, 
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Saturday 14 December 2024  18:00:05 -0500 (0:00:00.085)       0:05:52.914 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Saturday 14 December 2024  18:00:05 -0500 (0:00:00.069)       0:05:52.984 ***** 
ok: [managed-node2] => {
    "bytes": 10737418240, 
    "changed": false, 
    "lvm": "10g", 
    "parted": "10GiB", 
    "size": "10 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Saturday 14 December 2024  18:00:05 -0500 (0:00:00.386)       0:05:53.371 ***** 
skipping: [managed-node2] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Saturday 14 December 2024  18:00:06 -0500 (0:00:00.044)       0:05:53.415 ***** 
skipping: [managed-node2] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Saturday 14 December 2024  18:00:06 -0500 (0:00:00.051)       0:05:53.466 ***** 
skipping: [managed-node2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Saturday 14 December 2024  18:00:06 -0500 (0:00:00.073)       0:05:53.540 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Saturday 14 December 2024  18:00:06 -0500 (0:00:00.064)       0:05:53.605 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Saturday 14 December 2024  18:00:06 -0500 (0:00:00.054)       0:05:53.659 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Saturday 14 December 2024  18:00:06 -0500 (0:00:00.072)       0:05:53.732 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Saturday 14 December 2024  18:00:06 -0500 (0:00:00.073)       0:05:53.806 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Saturday 14 December 2024  18:00:06 -0500 (0:00:00.065)       0:05:53.871 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Saturday 14 December 2024  18:00:06 -0500 (0:00:00.071)       0:05:53.943 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Saturday 14 December 2024  18:00:06 -0500 (0:00:00.066)       0:05:54.009 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Saturday 14 December 2024  18:00:06 -0500 (0:00:00.058)       0:05:54.068 ***** 
skipping: [managed-node2] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Saturday 14 December 2024  18:00:06 -0500 (0:00:00.071)       0:05:54.140 ***** 
skipping: [managed-node2] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Saturday 14 December 2024  18:00:06 -0500 (0:00:00.057)       0:05:54.197 ***** 
skipping: [managed-node2] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Saturday 14 December 2024  18:00:06 -0500 (0:00:00.065)       0:05:54.263 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Saturday 14 December 2024  18:00:06 -0500 (0:00:00.070)       0:05:54.334 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Saturday 14 December 2024  18:00:07 -0500 (0:00:00.070)       0:05:54.404 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Saturday 14 December 2024  18:00:07 -0500 (0:00:00.082)       0:05:54.486 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Saturday 14 December 2024  18:00:07 -0500 (0:00:00.140)       0:05:54.627 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Saturday 14 December 2024  18:00:07 -0500 (0:00:00.059)       0:05:54.687 ***** 
ok: [managed-node2] => {
    "storage_test_actual_size": {
        "bytes": 4294967296, 
        "changed": false, 
        "failed": false, 
        "lvm": "4g", 
        "parted": "4GiB", 
        "size": "4 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Saturday 14 December 2024  18:00:07 -0500 (0:00:00.105)       0:05:54.793 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Saturday 14 December 2024  18:00:07 -0500 (0:00:00.074)       0:05:54.867 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Saturday 14 December 2024  18:00:07 -0500 (0:00:00.099)       0:05:54.967 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "lvs", 
        "--noheadings", 
        "--nameprefixes", 
        "--units=b", 
        "--nosuffix", 
        "--unquoted", 
        "-o", 
        "name,attr,cache_total_blocks,chunk_size,segtype", 
        "foo/test1"
    ], 
    "delta": "0:00:00.019880", 
    "end": "2024-12-14 18:00:08.053251", 
    "rc": 0, 
    "start": "2024-12-14 18:00:08.033371"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Saturday 14 December 2024  18:00:08 -0500 (0:00:00.575)       0:05:55.542 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    }, 
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Saturday 14 December 2024  18:00:08 -0500 (0:00:00.078)       0:05:55.621 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Saturday 14 December 2024  18:00:08 -0500 (0:00:00.089)       0:05:55.710 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Saturday 14 December 2024  18:00:08 -0500 (0:00:00.082)       0:05:55.793 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Saturday 14 December 2024  18:00:08 -0500 (0:00:00.067)       0:05:55.861 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Saturday 14 December 2024  18:00:08 -0500 (0:00:00.088)       0:05:55.949 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Saturday 14 December 2024  18:00:08 -0500 (0:00:00.071)       0:05:56.021 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    }, 
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Saturday 14 December 2024  18:00:08 -0500 (0:00:00.080)       0:05:56.102 ***** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Saturday 14 December 2024  18:00:08 -0500 (0:00:00.060)       0:05:56.162 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null, 
        "storage_test_crypttab": null, 
        "storage_test_fstab": null
    }, 
    "changed": false
}

TASK [Verify preservation of encryption settings on existing LVM volume] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:396
Saturday 14 December 2024  18:00:08 -0500 (0:00:00.053)       0:05:56.215 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  18:00:08 -0500 (0:00:00.135)       0:05:56.351 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  18:00:09 -0500 (0:00:00.089)       0:05:56.440 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  18:00:09 -0500 (0:00:00.074)       0:05:56.514 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  18:00:09 -0500 (0:00:00.143)       0:05:56.658 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  18:00:09 -0500 (0:00:00.102)       0:05:56.760 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  18:00:09 -0500 (0:00:00.104)       0:05:56.865 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  18:00:09 -0500 (0:00:00.056)       0:05:56.921 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  18:00:09 -0500 (0:00:00.059)       0:05:56.980 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  18:00:09 -0500 (0:00:00.243)       0:05:57.224 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  18:00:11 -0500 (0:00:01.390)       0:05:58.615 ***** 
ok: [managed-node2] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ], 
            "name": "foo", 
            "type": "lvm", 
            "volumes": [
                {
                    "mount_point": "/opt/test1", 
                    "name": "test1", 
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  18:00:11 -0500 (0:00:00.141)       0:05:58.756 ***** 
ok: [managed-node2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  18:00:11 -0500 (0:00:00.127)       0:05:58.884 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [
        "lvm2"
    ], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  18:00:15 -0500 (0:00:04.204)       0:06:03.088 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  18:00:15 -0500 (0:00:00.130)       0:06:03.219 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  18:00:15 -0500 (0:00:00.035)       0:06:03.255 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  18:00:15 -0500 (0:00:00.037)       0:06:03.293 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  18:00:15 -0500 (0:00:00.034)       0:06:03.327 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "7:lvm2-2.02.187-6.el7_9.5.x86_64 providing lvm2 is already installed", 
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: kpartx lvm2

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  18:00:16 -0500 (0:00:00.633)       0:06:03.961 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-cryptsetup@luks\\x2db6506dcf\\x2d75d2\\x2d4401\\x2d8040\\x2d0091a37b6dfd.service": {
                "name": "systemd-cryptsetup@luks\\x2db6506dcf\\x2d75d2\\x2d4401\\x2d8040\\x2d0091a37b6dfd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  18:00:17 -0500 (0:00:01.039)       0:06:05.000 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": [
            "systemd-cryptsetup@luks\\x2db6506dcf\\x2d75d2\\x2d4401\\x2d8040\\x2d0091a37b6dfd.service"
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  18:00:17 -0500 (0:00:00.066)       0:06:05.067 ***** 
changed: [managed-node2] => (item=systemd-cryptsetup@luks\x2db6506dcf\x2d75d2\x2d4401\x2d8040\x2d0091a37b6dfd.service) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": "systemd-cryptsetup@luks\\x2db6506dcf\\x2d75d2\\x2d4401\\x2d8040\\x2d0091a37b6dfd.service", 
    "name": "systemd-cryptsetup@luks\\x2db6506dcf\\x2d75d2\\x2d4401\\x2d8040\\x2d0091a37b6dfd.service", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "After": "systemd-readahead-collect.service cryptsetup-pre.target systemd-journald.socket dev-sda1.device system-systemd\\x2dcryptsetup.slice systemd-readahead-replay.service", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "Before": "cryptsetup.target umount.target", 
        "BindsTo": "dev-sda1.device", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "Conflicts": "umount.target", 
        "ControlPID": "0", 
        "DefaultDependencies": "no", 
        "Delegate": "no", 
        "Description": "Cryptography Setup for luks-b6506dcf-75d2-4401-8040-0091a37b6dfd", 
        "DevicePolicy": "auto", 
        "Documentation": "man:crypttab(5) man:systemd-cryptsetup-generator(8) man:systemd-cryptsetup@.service(8)", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-b6506dcf-75d2-4401-8040-0091a37b6dfd /dev/sda1 VALUE_SPECIFIED_IN_NO_LOG_PARAMETER  ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-b6506dcf-75d2-4401-8040-0091a37b6dfd ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2db6506dcf\\x2d75d2\\x2d4401\\x2d8040\\x2d0091a37b6dfd.service", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "systemd-cryptsetup@luks\\x2db6506dcf\\x2d75d2\\x2d4401\\x2d8040\\x2d0091a37b6dfd.service", 
        "IgnoreOnIsolate": "yes", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "4096", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "loaded", 
        "MainPID": "0", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "systemd-cryptsetup@luks\\x2db6506dcf\\x2d75d2\\x2d4401\\x2d8040\\x2d0091a37b6dfd.service", 
        "NeedDaemonReload": "yes", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "yes", 
        "RequiredBy": "cryptsetup.target", 
        "Requires": "system-systemd\\x2dcryptsetup.slice", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "Slice": "system-systemd\\x2dcryptsetup.slice", 
        "SourcePath": "/etc/crypttab", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "journal", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "0", 
        "TimeoutStopUSec": "0", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "Type": "oneshot", 
        "UMask": "0022", 
        "UnitFilePreset": "disabled", 
        "UnitFileState": "bad", 
        "WantedBy": "dev-sda1.device", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  18:00:18 -0500 (0:00:00.618)       0:06:05.685 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [
        "/dev/sdb", 
        "/dev/sdc", 
        "/dev/sdd", 
        "/dev/sde", 
        "/dev/sdf", 
        "/dev/sdg", 
        "/dev/sdh", 
        "/dev/sdi", 
        "/dev/xvda1", 
        "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c"
    ], 
    "mounts": [
        {
            "dump": 0, 
            "fstype": "xfs", 
            "group": null, 
            "mode": null, 
            "opts": "defaults", 
            "owner": null, 
            "passno": 0, 
            "path": "/opt/test1", 
            "src": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
            "state": "mounted"
        }
    ], 
    "packages": [
        "cryptsetup", 
        "xfsprogs", 
        "e2fsprogs", 
        "lvm2"
    ], 
    "pools": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_cipher": null, 
            "encryption_clevis_pin": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": null, 
            "encryption_tang_thumbprint": null, 
            "encryption_tang_url": null, 
            "grow_to_fill": false, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "shared": false, 
            "state": "present", 
            "type": "lvm", 
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                    "_kernel_device": "/dev/dm-1", 
                    "_mount_id": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                    "_raw_device": "/dev/mapper/foo-test1", 
                    "_raw_kernel_device": "/dev/dm-0", 
                    "cache_devices": [], 
                    "cache_mode": null, 
                    "cache_size": 0, 
                    "cached": false, 
                    "compression": null, 
                    "deduplication": null, 
                    "disks": [
                        "sda"
                    ], 
                    "encryption": true, 
                    "encryption_cipher": null, 
                    "encryption_key": null, 
                    "encryption_key_size": 0, 
                    "encryption_luks_version": "luks1", 
                    "encryption_password": null, 
                    "fs_create_options": "", 
                    "fs_label": "", 
                    "fs_overwrite_existing": true, 
                    "fs_type": "xfs", 
                    "mount_check": 0, 
                    "mount_device_identifier": "uuid", 
                    "mount_group": null, 
                    "mount_mode": null, 
                    "mount_options": "defaults", 
                    "mount_passno": 0, 
                    "mount_point": "/opt/test1", 
                    "mount_user": null, 
                    "name": "test1", 
                    "raid_chunk_size": null, 
                    "raid_device_count": null, 
                    "raid_disks": [], 
                    "raid_level": null, 
                    "raid_metadata_version": null, 
                    "raid_spare_count": null, 
                    "raid_stripe_size": null, 
                    "size": "4g", 
                    "state": "present", 
                    "thin": false, 
                    "thin_pool_name": null, 
                    "thin_pool_size": null, 
                    "type": "lvm", 
                    "vdo_pool_size": null
                }
            ]
        }
    ], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Saturday 14 December 2024  18:00:22 -0500 (0:00:04.147)       0:06:09.833 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Saturday 14 December 2024  18:00:22 -0500 (0:00:00.046)       0:06:09.879 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217188.1247318, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "7e4e8b3ad4d6e63ef5f280d635ca488627b2dd22", 
        "ctime": 1734217188.1217318, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 263588, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0644", 
        "mtime": 1734217188.1217318, 
        "nlink": 1, 
        "path": "/etc/fstab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 1311, 
        "uid": 0, 
        "version": "606402765", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Saturday 14 December 2024  18:00:22 -0500 (0:00:00.443)       0:06:10.322 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  18:00:23 -0500 (0:00:00.083)       0:06:10.406 ***** 
changed: [managed-node2] => (item=systemd-cryptsetup@luks\x2db6506dcf\x2d75d2\x2d4401\x2d8040\x2d0091a37b6dfd.service) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": "systemd-cryptsetup@luks\\x2db6506dcf\\x2d75d2\\x2d4401\\x2d8040\\x2d0091a37b6dfd.service", 
    "name": "systemd-cryptsetup@luks\\x2db6506dcf\\x2d75d2\\x2d4401\\x2d8040\\x2d0091a37b6dfd.service", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "ControlPID": "0", 
        "DefaultDependencies": "yes", 
        "Delegate": "no", 
        "Description": "systemd-cryptsetup@luks\\x2db6506dcf\\x2d75d2\\x2d4401\\x2d8040\\x2d0091a37b6dfd.service", 
        "DevicePolicy": "auto", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/dev/null", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "systemd-cryptsetup@luks\\x2db6506dcf\\x2d75d2\\x2d4401\\x2d8040\\x2d0091a37b6dfd.service", 
        "IgnoreOnIsolate": "no", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "1048576", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "masked", 
        "MainPID": "0", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "systemd-cryptsetup@luks\\x2db6506dcf\\x2d75d2\\x2d4401\\x2d8040\\x2d0091a37b6dfd.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "no", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "inherit", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "1min 30s", 
        "TimeoutStopUSec": "1min 30s", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "UMask": "0022", 
        "UnitFileState": "bad", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Saturday 14 December 2024  18:00:23 -0500 (0:00:00.631)       0:06:11.037 ***** 
ok: [managed-node2] => {
    "blivet_output": {
        "actions": [], 
        "changed": false, 
        "crypts": [], 
        "failed": false, 
        "leaves": [
            "/dev/sdb", 
            "/dev/sdc", 
            "/dev/sdd", 
            "/dev/sde", 
            "/dev/sdf", 
            "/dev/sdg", 
            "/dev/sdh", 
            "/dev/sdi", 
            "/dev/xvda1", 
            "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c"
        ], 
        "mounts": [
            {
                "dump": 0, 
                "fstype": "xfs", 
                "group": null, 
                "mode": null, 
                "opts": "defaults", 
                "owner": null, 
                "passno": 0, 
                "path": "/opt/test1", 
                "src": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                "state": "mounted"
            }
        ], 
        "packages": [
            "cryptsetup", 
            "xfsprogs", 
            "e2fsprogs", 
            "lvm2"
        ], 
        "pools": [
            {
                "disks": [
                    "sda"
                ], 
                "encryption": false, 
                "encryption_cipher": null, 
                "encryption_clevis_pin": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": null, 
                "encryption_tang_thumbprint": null, 
                "encryption_tang_url": null, 
                "grow_to_fill": false, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "shared": false, 
                "state": "present", 
                "type": "lvm", 
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                        "_kernel_device": "/dev/dm-1", 
                        "_mount_id": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                        "_raw_device": "/dev/mapper/foo-test1", 
                        "_raw_kernel_device": "/dev/dm-0", 
                        "cache_devices": [], 
                        "cache_mode": null, 
                        "cache_size": 0, 
                        "cached": false, 
                        "compression": null, 
                        "deduplication": null, 
                        "disks": [
                            "sda"
                        ], 
                        "encryption": true, 
                        "encryption_cipher": null, 
                        "encryption_key": null, 
                        "encryption_key_size": 0, 
                        "encryption_luks_version": "luks1", 
                        "encryption_password": null, 
                        "fs_create_options": "", 
                        "fs_label": "", 
                        "fs_overwrite_existing": true, 
                        "fs_type": "xfs", 
                        "mount_check": 0, 
                        "mount_device_identifier": "uuid", 
                        "mount_group": null, 
                        "mount_mode": null, 
                        "mount_options": "defaults", 
                        "mount_passno": 0, 
                        "mount_point": "/opt/test1", 
                        "mount_user": null, 
                        "name": "test1", 
                        "raid_chunk_size": null, 
                        "raid_device_count": null, 
                        "raid_disks": [], 
                        "raid_level": null, 
                        "raid_metadata_version": null, 
                        "raid_spare_count": null, 
                        "raid_stripe_size": null, 
                        "size": "4g", 
                        "state": "present", 
                        "thin": false, 
                        "thin_pool_name": null, 
                        "thin_pool_size": null, 
                        "type": "lvm", 
                        "vdo_pool_size": null
                    }
                ]
            }
        ], 
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Saturday 14 December 2024  18:00:23 -0500 (0:00:00.096)       0:06:11.133 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ], 
                "encryption": false, 
                "encryption_cipher": null, 
                "encryption_clevis_pin": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": null, 
                "encryption_tang_thumbprint": null, 
                "encryption_tang_url": null, 
                "grow_to_fill": false, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "shared": false, 
                "state": "present", 
                "type": "lvm", 
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                        "_kernel_device": "/dev/dm-1", 
                        "_mount_id": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                        "_raw_device": "/dev/mapper/foo-test1", 
                        "_raw_kernel_device": "/dev/dm-0", 
                        "cache_devices": [], 
                        "cache_mode": null, 
                        "cache_size": 0, 
                        "cached": false, 
                        "compression": null, 
                        "deduplication": null, 
                        "disks": [
                            "sda"
                        ], 
                        "encryption": true, 
                        "encryption_cipher": null, 
                        "encryption_key": null, 
                        "encryption_key_size": 0, 
                        "encryption_luks_version": "luks1", 
                        "encryption_password": null, 
                        "fs_create_options": "", 
                        "fs_label": "", 
                        "fs_overwrite_existing": true, 
                        "fs_type": "xfs", 
                        "mount_check": 0, 
                        "mount_device_identifier": "uuid", 
                        "mount_group": null, 
                        "mount_mode": null, 
                        "mount_options": "defaults", 
                        "mount_passno": 0, 
                        "mount_point": "/opt/test1", 
                        "mount_user": null, 
                        "name": "test1", 
                        "raid_chunk_size": null, 
                        "raid_device_count": null, 
                        "raid_disks": [], 
                        "raid_level": null, 
                        "raid_metadata_version": null, 
                        "raid_spare_count": null, 
                        "raid_stripe_size": null, 
                        "size": "4g", 
                        "state": "present", 
                        "thin": false, 
                        "thin_pool_name": null, 
                        "thin_pool_size": null, 
                        "type": "lvm", 
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Saturday 14 December 2024  18:00:23 -0500 (0:00:00.070)       0:06:11.203 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Saturday 14 December 2024  18:00:23 -0500 (0:00:00.060)       0:06:11.264 ***** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Saturday 14 December 2024  18:00:23 -0500 (0:00:00.066)       0:06:11.331 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Saturday 14 December 2024  18:00:24 -0500 (0:00:00.576)       0:06:11.907 ***** 
ok: [managed-node2] => (item={u'src': u'/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": false, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
        "state": "mounted"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Saturday 14 December 2024  18:00:24 -0500 (0:00:00.457)       0:06:12.365 ***** 
skipping: [managed-node2] => (item={u'src': u'/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'})  => {
    "ansible_loop_var": "mount_info", 
    "changed": false, 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
        "state": "mounted"
    }, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Saturday 14 December 2024  18:00:25 -0500 (0:00:00.082)       0:06:12.447 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Saturday 14 December 2024  18:00:25 -0500 (0:00:00.591)       0:06:13.038 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217191.81573, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "184e4520d6a02c13cdc2844673950a75c9ab85fd", 
        "ctime": 1734217189.7647312, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 917511, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0600", 
        "mtime": 1734217189.7647312, 
        "nlink": 1, 
        "path": "/etc/crypttab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": false, 
        "roth": false, 
        "rusr": true, 
        "size": 66, 
        "uid": 0, 
        "version": "606409565", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Saturday 14 December 2024  18:00:26 -0500 (0:00:00.456)       0:06:13.494 ***** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Saturday 14 December 2024  18:00:26 -0500 (0:00:00.055)       0:06:13.550 ***** 
ok: [managed-node2]

TASK [Assert preservation of encryption settings on existing LVM volume] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:410
Saturday 14 December 2024  18:00:28 -0500 (0:00:02.077)       0:06:15.628 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify role results] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:417
Saturday 14 December 2024  18:00:28 -0500 (0:00:00.137)       0:06:15.766 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node2

TASK [Print out pool information] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Saturday 14 December 2024  18:00:28 -0500 (0:00:00.256)       0:06:16.022 ***** 
ok: [managed-node2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_cipher": null, 
            "encryption_clevis_pin": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": null, 
            "encryption_tang_thumbprint": null, 
            "encryption_tang_url": null, 
            "grow_to_fill": false, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "shared": false, 
            "state": "present", 
            "type": "lvm", 
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                    "_kernel_device": "/dev/dm-1", 
                    "_mount_id": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                    "_raw_device": "/dev/mapper/foo-test1", 
                    "_raw_kernel_device": "/dev/dm-0", 
                    "cache_devices": [], 
                    "cache_mode": null, 
                    "cache_size": 0, 
                    "cached": false, 
                    "compression": null, 
                    "deduplication": null, 
                    "disks": [
                        "sda"
                    ], 
                    "encryption": true, 
                    "encryption_cipher": null, 
                    "encryption_key": null, 
                    "encryption_key_size": 0, 
                    "encryption_luks_version": "luks1", 
                    "encryption_password": null, 
                    "fs_create_options": "", 
                    "fs_label": "", 
                    "fs_overwrite_existing": true, 
                    "fs_type": "xfs", 
                    "mount_check": 0, 
                    "mount_device_identifier": "uuid", 
                    "mount_group": null, 
                    "mount_mode": null, 
                    "mount_options": "defaults", 
                    "mount_passno": 0, 
                    "mount_point": "/opt/test1", 
                    "mount_user": null, 
                    "name": "test1", 
                    "raid_chunk_size": null, 
                    "raid_device_count": null, 
                    "raid_disks": [], 
                    "raid_level": null, 
                    "raid_metadata_version": null, 
                    "raid_spare_count": null, 
                    "raid_stripe_size": null, 
                    "size": "4g", 
                    "state": "present", 
                    "thin": false, 
                    "thin_pool_name": null, 
                    "thin_pool_size": null, 
                    "type": "lvm", 
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Saturday 14 December 2024  18:00:28 -0500 (0:00:00.157)       0:06:16.180 ***** 
skipping: [managed-node2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Saturday 14 December 2024  18:00:28 -0500 (0:00:00.082)       0:06:16.263 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "crypto_LUKS", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/mapper/foo-test1", 
            "size": "4G", 
            "type": "lvm", 
            "uuid": "3365bc93-2fc0-47e8-b784-b63a2126f78c"
        }, 
        "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c": {
            "fstype": "xfs", 
            "label": "", 
            "mountpoint": "/opt/test1", 
            "name": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
            "size": "4G", 
            "type": "crypt", 
            "uuid": "4f26c9ff-d55a-4463-b590-1cd68a73451b"
        }, 
        "/dev/sda": {
            "fstype": "LVM2_member", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sda", 
            "size": "10G", 
            "type": "disk", 
            "uuid": "xa7q16-X4RQ-vFWY-Ftcj-AFkA-0I7B-z0LDg5"
        }, 
        "/dev/sdb": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdb", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdc": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdc", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdd": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdd", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sde": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sde", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdf": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdf", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdg": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdg", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdh": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdh", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdi": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdi", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/xvda", 
            "size": "250G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda1": {
            "fstype": "ext4", 
            "label": "", 
            "mountpoint": "/", 
            "name": "/dev/xvda1", 
            "size": "250G", 
            "type": "partition", 
            "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Saturday 14 December 2024  18:00:29 -0500 (0:00:00.576)       0:06:16.839 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/fstab"
    ], 
    "delta": "0:00:00.003786", 
    "end": "2024-12-14 18:00:30.021952", 
    "rc": 0, 
    "start": "2024-12-14 18:00:30.018166"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Thu Jun 20 10:23:46 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 /                       ext4    defaults        1 1
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Saturday 14 December 2024  18:00:30 -0500 (0:00:00.695)       0:06:17.535 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/crypttab"
    ], 
    "delta": "0:00:00.003206", 
    "end": "2024-12-14 18:00:30.537475", 
    "failed_when_result": false, 
    "rc": 0, 
    "start": "2024-12-14 18:00:30.534269"
}

STDOUT:

luks-3365bc93-2fc0-47e8-b784-b63a2126f78c /dev/mapper/foo-test1 -

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Saturday 14 December 2024  18:00:30 -0500 (0:00:00.474)       0:06:18.009 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node2

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Saturday 14 December 2024  18:00:30 -0500 (0:00:00.141)       0:06:18.151 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members", 
            "volumes"
        ]
    }, 
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Saturday 14 December 2024  18:00:30 -0500 (0:00:00.087)       0:06:18.239 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "vgs", 
        "--noheadings", 
        "--binary", 
        "-o", 
        "shared", 
        "foo"
    ], 
    "delta": "0:00:00.018123", 
    "end": "2024-12-14 18:00:31.311671", 
    "rc": 0, 
    "start": "2024-12-14 18:00:31.293548"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Saturday 14 December 2024  18:00:31 -0500 (0:00:00.632)       0:06:18.872 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Saturday 14 December 2024  18:00:31 -0500 (0:00:00.152)       0:06:19.024 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node2

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Saturday 14 December 2024  18:00:31 -0500 (0:00:00.176)       0:06:19.201 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1", 
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    }, 
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Saturday 14 December 2024  18:00:31 -0500 (0:00:00.090)       0:06:19.292 ***** 
ok: [managed-node2] => (item=/dev/sda) => {
    "ansible_loop_var": "pv", 
    "changed": false, 
    "device": "/dev/sda", 
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Saturday 14 December 2024  18:00:32 -0500 (0:00:00.421)       0:06:19.713 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    }, 
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Saturday 14 December 2024  18:00:32 -0500 (0:00:00.087)       0:06:19.801 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    }, 
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Saturday 14 December 2024  18:00:32 -0500 (0:00:00.074)       0:06:19.876 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Saturday 14 December 2024  18:00:32 -0500 (0:00:00.083)       0:06:19.960 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    }, 
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Saturday 14 December 2024  18:00:32 -0500 (0:00:00.074)       0:06:20.034 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    }, 
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Saturday 14 December 2024  18:00:32 -0500 (0:00:00.104)       0:06:20.139 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Saturday 14 December 2024  18:00:32 -0500 (0:00:00.100)       0:06:20.239 ***** 
ok: [managed-node2] => (item=/dev/sda) => {
    "ansible_loop_var": "pv", 
    "changed": false, 
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Saturday 14 December 2024  18:00:32 -0500 (0:00:00.133)       0:06:20.373 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.8.171 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Saturday 14 December 2024  18:00:33 -0500 (0:00:00.478)       0:06:20.852 ***** 
skipping: [managed-node2] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv", 
    "changed": false, 
    "skip_reason": "Conditional result was False", 
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Saturday 14 December 2024  18:00:33 -0500 (0:00:00.077)       0:06:20.930 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node2

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Saturday 14 December 2024  18:00:33 -0500 (0:00:00.122)       0:06:21.052 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Saturday 14 December 2024  18:00:33 -0500 (0:00:00.062)       0:06:21.115 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Saturday 14 December 2024  18:00:33 -0500 (0:00:00.051)       0:06:21.166 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Saturday 14 December 2024  18:00:33 -0500 (0:00:00.042)       0:06:21.208 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Saturday 14 December 2024  18:00:33 -0500 (0:00:00.037)       0:06:21.245 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Saturday 14 December 2024  18:00:33 -0500 (0:00:00.039)       0:06:21.285 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Saturday 14 December 2024  18:00:33 -0500 (0:00:00.042)       0:06:21.328 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Saturday 14 December 2024  18:00:33 -0500 (0:00:00.052)       0:06:21.380 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Saturday 14 December 2024  18:00:34 -0500 (0:00:00.077)       0:06:21.457 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Saturday 14 December 2024  18:00:34 -0500 (0:00:00.061)       0:06:21.519 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Saturday 14 December 2024  18:00:34 -0500 (0:00:00.102)       0:06:21.621 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null, 
        "storage_test_md_chunk_size_re": null, 
        "storage_test_md_metadata_version_re": null, 
        "storage_test_md_spare_devices_re": null
    }, 
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Saturday 14 December 2024  18:00:34 -0500 (0:00:00.086)       0:06:21.708 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Saturday 14 December 2024  18:00:34 -0500 (0:00:00.212)       0:06:21.920 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node2

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Saturday 14 December 2024  18:00:34 -0500 (0:00:00.141)       0:06:22.062 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Saturday 14 December 2024  18:00:34 -0500 (0:00:00.066)       0:06:22.129 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Saturday 14 December 2024  18:00:34 -0500 (0:00:00.059)       0:06:22.188 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Saturday 14 December 2024  18:00:34 -0500 (0:00:00.067)       0:06:22.256 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Saturday 14 December 2024  18:00:34 -0500 (0:00:00.060)       0:06:22.316 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Saturday 14 December 2024  18:00:34 -0500 (0:00:00.065)       0:06:22.381 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Saturday 14 December 2024  18:00:35 -0500 (0:00:00.091)       0:06:22.473 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Saturday 14 December 2024  18:00:35 -0500 (0:00:00.082)       0:06:22.556 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Saturday 14 December 2024  18:00:35 -0500 (0:00:00.179)       0:06:22.736 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node2

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Saturday 14 December 2024  18:00:35 -0500 (0:00:00.222)       0:06:22.958 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Saturday 14 December 2024  18:00:35 -0500 (0:00:00.075)       0:06:23.033 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Saturday 14 December 2024  18:00:35 -0500 (0:00:00.058)       0:06:23.092 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Saturday 14 December 2024  18:00:35 -0500 (0:00:00.063)       0:06:23.156 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    }, 
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Saturday 14 December 2024  18:00:35 -0500 (0:00:00.057)       0:06:23.213 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node2

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Saturday 14 December 2024  18:00:36 -0500 (0:00:00.237)       0:06:23.451 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0", 
        "_storage_test_expected_crypttab_key_file": "-"
    }, 
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Saturday 14 December 2024  18:00:36 -0500 (0:00:00.076)       0:06:23.527 ***** 
skipping: [managed-node2] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda", 
    "ansible_loop_var": "_storage_test_pool_member_path", 
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Saturday 14 December 2024  18:00:36 -0500 (0:00:00.075)       0:06:23.603 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node2

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Saturday 14 December 2024  18:00:36 -0500 (0:00:00.114)       0:06:23.717 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    }, 
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Saturday 14 December 2024  18:00:36 -0500 (0:00:00.065)       0:06:23.782 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Saturday 14 December 2024  18:00:36 -0500 (0:00:00.075)       0:06:23.858 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Saturday 14 December 2024  18:00:36 -0500 (0:00:00.086)       0:06:23.944 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Saturday 14 December 2024  18:00:36 -0500 (0:00:00.082)       0:06:24.026 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Saturday 14 December 2024  18:00:36 -0500 (0:00:00.055)       0:06:24.081 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    }, 
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Saturday 14 December 2024  18:00:36 -0500 (0:00:00.061)       0:06:24.143 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null, 
        "_storage_test_crypttab_key_file": null
    }, 
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Saturday 14 December 2024  18:00:36 -0500 (0:00:00.056)       0:06:24.199 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Saturday 14 December 2024  18:00:36 -0500 (0:00:00.150)       0:06:24.350 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node2

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Saturday 14 December 2024  18:00:37 -0500 (0:00:00.176)       0:06:24.527 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Saturday 14 December 2024  18:00:37 -0500 (0:00:00.079)       0:06:24.606 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Saturday 14 December 2024  18:00:37 -0500 (0:00:00.071)       0:06:24.678 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Saturday 14 December 2024  18:00:37 -0500 (0:00:00.063)       0:06:24.741 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Saturday 14 December 2024  18:00:37 -0500 (0:00:00.067)       0:06:24.808 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Saturday 14 December 2024  18:00:37 -0500 (0:00:00.063)       0:06:24.871 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Saturday 14 December 2024  18:00:37 -0500 (0:00:00.058)       0:06:24.930 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    }, 
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Saturday 14 December 2024  18:00:37 -0500 (0:00:00.064)       0:06:24.995 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node2

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Saturday 14 December 2024  18:00:37 -0500 (0:00:00.223)       0:06:25.218 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Saturday 14 December 2024  18:00:37 -0500 (0:00:00.080)       0:06:25.298 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Saturday 14 December 2024  18:00:38 -0500 (0:00:00.117)       0:06:25.416 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Saturday 14 December 2024  18:00:38 -0500 (0:00:00.081)       0:06:25.497 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Saturday 14 December 2024  18:00:38 -0500 (0:00:00.061)       0:06:25.558 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Saturday 14 December 2024  18:00:38 -0500 (0:00:00.057)       0:06:25.616 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    }, 
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Saturday 14 December 2024  18:00:38 -0500 (0:00:00.080)       0:06:25.697 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null, 
        "_storage_test_expected_pv_count": null, 
        "_storage_test_expected_pv_type": null, 
        "_storage_test_pool_pvs": [], 
        "_storage_test_pool_pvs_lvm": []
    }, 
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Saturday 14 December 2024  18:00:38 -0500 (0:00:00.061)       0:06:25.758 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node2

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Saturday 14 December 2024  18:00:38 -0500 (0:00:00.120)       0:06:25.879 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true, 
        "_storage_volume_tests": [
            "mount", 
            "fstab", 
            "fs", 
            "device", 
            "encryption", 
            "md", 
            "size", 
            "cache"
        ]
    }, 
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Saturday 14 December 2024  18:00:38 -0500 (0:00:00.099)       0:06:25.979 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node2

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Saturday 14 December 2024  18:00:38 -0500 (0:00:00.371)       0:06:26.350 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c"
    }, 
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Saturday 14 December 2024  18:00:39 -0500 (0:00:00.105)       0:06:26.456 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1", 
        "storage_test_swap_expected_matches": "0"
    }, 
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Saturday 14 December 2024  18:00:39 -0500 (0:00:00.141)       0:06:26.597 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Saturday 14 December 2024  18:00:39 -0500 (0:00:00.074)       0:06:26.672 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Saturday 14 December 2024  18:00:39 -0500 (0:00:00.132)       0:06:26.805 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Saturday 14 December 2024  18:00:39 -0500 (0:00:00.138)       0:06:26.944 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Saturday 14 December 2024  18:00:39 -0500 (0:00:00.124)       0:06:27.069 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Saturday 14 December 2024  18:00:39 -0500 (0:00:00.090)       0:06:27.159 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Saturday 14 December 2024  18:00:39 -0500 (0:00:00.072)       0:06:27.232 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Saturday 14 December 2024  18:00:39 -0500 (0:00:00.093)       0:06:27.325 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Saturday 14 December 2024  18:00:40 -0500 (0:00:00.105)       0:06:27.431 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null, 
        "storage_test_mount_expected_mount_point": null, 
        "storage_test_swap_expected_matches": null, 
        "storage_test_swaps": null, 
        "storage_test_sys_node": null
    }, 
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Saturday 14 December 2024  18:00:40 -0500 (0:00:00.121)       0:06:27.553 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1", 
        "storage_test_fstab_expected_mount_options_matches": "1", 
        "storage_test_fstab_expected_mount_point_matches": "1", 
        "storage_test_fstab_id_matches": [
            "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c "
        ], 
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ], 
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    }, 
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Saturday 14 December 2024  18:00:40 -0500 (0:00:00.147)       0:06:27.701 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Saturday 14 December 2024  18:00:40 -0500 (0:00:00.135)       0:06:27.837 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Saturday 14 December 2024  18:00:40 -0500 (0:00:00.098)       0:06:27.935 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Saturday 14 December 2024  18:00:40 -0500 (0:00:00.075)       0:06:28.011 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Saturday 14 December 2024  18:00:40 -0500 (0:00:00.077)       0:06:28.088 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null, 
        "storage_test_fstab_expected_mount_options_matches": null, 
        "storage_test_fstab_expected_mount_point_matches": null, 
        "storage_test_fstab_id_matches": null, 
        "storage_test_fstab_mount_options_matches": null, 
        "storage_test_fstab_mount_point_matches": null
    }, 
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Saturday 14 December 2024  18:00:40 -0500 (0:00:00.086)       0:06:28.175 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Saturday 14 December 2024  18:00:40 -0500 (0:00:00.096)       0:06:28.272 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Saturday 14 December 2024  18:00:41 -0500 (0:00:00.166)       0:06:28.438 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217202.2067254, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "ctime": 1734217185.683733, 
        "dev": 5, 
        "device_type": 64768, 
        "executable": false, 
        "exists": true, 
        "gid": 6, 
        "gr_name": "disk", 
        "inode": 431513, 
        "isblk": true, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": false, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/symlink", 
        "mode": "0660", 
        "mtime": 1734217185.683733, 
        "nlink": 1, 
        "path": "/dev/mapper/foo-test1", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": null, 
        "wgrp": true, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Saturday 14 December 2024  18:00:41 -0500 (0:00:00.436)       0:06:28.875 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Saturday 14 December 2024  18:00:41 -0500 (0:00:00.074)       0:06:28.949 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Saturday 14 December 2024  18:00:41 -0500 (0:00:00.059)       0:06:29.009 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Saturday 14 December 2024  18:00:41 -0500 (0:00:00.076)       0:06:29.085 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    }, 
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Saturday 14 December 2024  18:00:41 -0500 (0:00:00.067)       0:06:29.153 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Saturday 14 December 2024  18:00:41 -0500 (0:00:00.061)       0:06:29.214 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Saturday 14 December 2024  18:00:41 -0500 (0:00:00.070)       0:06:29.284 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217185.796733, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "ctime": 1734217185.796733, 
        "dev": 5, 
        "device_type": 64769, 
        "executable": false, 
        "exists": true, 
        "gid": 6, 
        "gr_name": "disk", 
        "inode": 431588, 
        "isblk": true, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": false, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/symlink", 
        "mode": "0660", 
        "mtime": 1734217185.796733, 
        "nlink": 1, 
        "path": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": null, 
        "wgrp": true, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Saturday 14 December 2024  18:00:42 -0500 (0:00:00.433)       0:06:29.717 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}
lsrpackages: cryptsetup

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Saturday 14 December 2024  18:00:43 -0500 (0:00:00.722)       0:06:30.440 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cryptsetup", 
        "luksDump", 
        "/dev/mapper/foo-test1"
    ], 
    "delta": "0:00:00.026610", 
    "end": "2024-12-14 18:00:43.371607", 
    "rc": 0, 
    "start": "2024-12-14 18:00:43.344997"
}

STDOUT:

LUKS header information for /dev/mapper/foo-test1

Version:       	1
Cipher name:   	aes
Cipher mode:   	xts-plain64
Hash spec:     	sha256
Payload offset:	8192
MK bits:       	512
MK digest:     	6a ba c5 ed 79 fa 96 8d 6d d0 43 0a e8 5a 19 5c 6b 61 81 cd 
MK salt:       	40 98 5d e4 16 55 f3 9b 15 76 b2 1a b3 93 d7 4b 
               	f7 e0 a6 ec c0 1a ec 43 d0 ab 5a a5 41 ea b8 8b 
MK iterations: 	23272
UUID:          	3365bc93-2fc0-47e8-b784-b63a2126f78c

Key Slot 0: ENABLED
	Iterations:         	372362
	Salt:               	32 49 4d 0e a0 ac 4e be e7 c0 da 34 86 ff b4 98 
	                      	d4 e7 55 37 99 03 47 7d f9 b2 fb 14 94 4a a7 f1 
	Key material offset:	8
	AF stripes:            	4000
Key Slot 1: DISABLED
Key Slot 2: DISABLED
Key Slot 3: DISABLED
Key Slot 4: DISABLED
Key Slot 5: DISABLED
Key Slot 6: DISABLED
Key Slot 7: DISABLED

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Saturday 14 December 2024  18:00:43 -0500 (0:00:00.405)       0:06:30.845 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Saturday 14 December 2024  18:00:43 -0500 (0:00:00.076)       0:06:30.922 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Saturday 14 December 2024  18:00:43 -0500 (0:00:00.080)       0:06:31.003 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Saturday 14 December 2024  18:00:43 -0500 (0:00:00.078)       0:06:31.081 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Saturday 14 December 2024  18:00:43 -0500 (0:00:00.073)       0:06:31.155 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Saturday 14 December 2024  18:00:43 -0500 (0:00:00.081)       0:06:31.237 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Saturday 14 December 2024  18:00:43 -0500 (0:00:00.062)       0:06:31.299 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Saturday 14 December 2024  18:00:43 -0500 (0:00:00.041)       0:06:31.341 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [
            "luks-3365bc93-2fc0-47e8-b784-b63a2126f78c /dev/mapper/foo-test1 -"
        ], 
        "_storage_test_expected_crypttab_entries": "1", 
        "_storage_test_expected_crypttab_key_file": "-"
    }, 
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Saturday 14 December 2024  18:00:44 -0500 (0:00:00.057)       0:06:31.398 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Saturday 14 December 2024  18:00:44 -0500 (0:00:00.050)       0:06:31.449 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Saturday 14 December 2024  18:00:44 -0500 (0:00:00.048)       0:06:31.498 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Saturday 14 December 2024  18:00:44 -0500 (0:00:00.048)       0:06:31.547 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Saturday 14 December 2024  18:00:44 -0500 (0:00:00.053)       0:06:31.600 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null, 
        "_storage_test_expected_crypttab_entries": null, 
        "_storage_test_expected_crypttab_key_file": null
    }, 
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Saturday 14 December 2024  18:00:44 -0500 (0:00:00.048)       0:06:31.649 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Saturday 14 December 2024  18:00:44 -0500 (0:00:00.059)       0:06:31.708 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Saturday 14 December 2024  18:00:44 -0500 (0:00:00.060)       0:06:31.769 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Saturday 14 December 2024  18:00:44 -0500 (0:00:00.058)       0:06:31.827 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Saturday 14 December 2024  18:00:44 -0500 (0:00:00.060)       0:06:31.887 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Saturday 14 December 2024  18:00:44 -0500 (0:00:00.061)       0:06:31.949 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Saturday 14 December 2024  18:00:44 -0500 (0:00:00.059)       0:06:32.008 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Saturday 14 December 2024  18:00:44 -0500 (0:00:00.059)       0:06:32.068 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Saturday 14 December 2024  18:00:44 -0500 (0:00:00.059)       0:06:32.127 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Saturday 14 December 2024  18:00:44 -0500 (0:00:00.062)       0:06:32.190 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Saturday 14 December 2024  18:00:44 -0500 (0:00:00.060)       0:06:32.250 ***** 
ok: [managed-node2] => {
    "bytes": 4294967296, 
    "changed": false, 
    "lvm": "4g", 
    "parted": "4GiB", 
    "size": "4 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Saturday 14 December 2024  18:00:45 -0500 (0:00:00.435)       0:06:32.686 ***** 
ok: [managed-node2] => {
    "bytes": 4294967296, 
    "changed": false, 
    "lvm": "4g", 
    "parted": "4GiB", 
    "size": "4 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Saturday 14 December 2024  18:00:45 -0500 (0:00:00.394)       0:06:33.081 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_expected_size": "4294967296"
    }, 
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Saturday 14 December 2024  18:00:45 -0500 (0:00:00.054)       0:06:33.136 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Saturday 14 December 2024  18:00:45 -0500 (0:00:00.042)       0:06:33.178 ***** 
ok: [managed-node2] => {
    "bytes": 10737418240, 
    "changed": false, 
    "lvm": "10g", 
    "parted": "10GiB", 
    "size": "10 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Saturday 14 December 2024  18:00:46 -0500 (0:00:00.375)       0:06:33.553 ***** 
skipping: [managed-node2] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Saturday 14 December 2024  18:00:46 -0500 (0:00:00.056)       0:06:33.610 ***** 
skipping: [managed-node2] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Saturday 14 December 2024  18:00:46 -0500 (0:00:00.074)       0:06:33.685 ***** 
skipping: [managed-node2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Saturday 14 December 2024  18:00:46 -0500 (0:00:00.074)       0:06:33.759 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Saturday 14 December 2024  18:00:46 -0500 (0:00:00.063)       0:06:33.823 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Saturday 14 December 2024  18:00:46 -0500 (0:00:00.061)       0:06:33.884 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Saturday 14 December 2024  18:00:46 -0500 (0:00:00.061)       0:06:33.945 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Saturday 14 December 2024  18:00:46 -0500 (0:00:00.074)       0:06:34.020 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Saturday 14 December 2024  18:00:46 -0500 (0:00:00.060)       0:06:34.080 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Saturday 14 December 2024  18:00:46 -0500 (0:00:00.047)       0:06:34.128 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Saturday 14 December 2024  18:00:46 -0500 (0:00:00.044)       0:06:34.173 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Saturday 14 December 2024  18:00:46 -0500 (0:00:00.046)       0:06:34.219 ***** 
skipping: [managed-node2] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Saturday 14 December 2024  18:00:46 -0500 (0:00:00.044)       0:06:34.264 ***** 
skipping: [managed-node2] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Saturday 14 December 2024  18:00:46 -0500 (0:00:00.051)       0:06:34.315 ***** 
skipping: [managed-node2] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Saturday 14 December 2024  18:00:47 -0500 (0:00:00.068)       0:06:34.384 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Saturday 14 December 2024  18:00:47 -0500 (0:00:00.061)       0:06:34.446 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Saturday 14 December 2024  18:00:47 -0500 (0:00:00.056)       0:06:34.502 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Saturday 14 December 2024  18:00:47 -0500 (0:00:00.080)       0:06:34.583 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Saturday 14 December 2024  18:00:47 -0500 (0:00:00.056)       0:06:34.640 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Saturday 14 December 2024  18:00:47 -0500 (0:00:00.089)       0:06:34.730 ***** 
ok: [managed-node2] => {
    "storage_test_actual_size": {
        "bytes": 4294967296, 
        "changed": false, 
        "failed": false, 
        "lvm": "4g", 
        "parted": "4GiB", 
        "size": "4 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Saturday 14 December 2024  18:00:47 -0500 (0:00:00.114)       0:06:34.844 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Saturday 14 December 2024  18:00:47 -0500 (0:00:00.116)       0:06:34.961 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Saturday 14 December 2024  18:00:47 -0500 (0:00:00.091)       0:06:35.052 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "lvs", 
        "--noheadings", 
        "--nameprefixes", 
        "--units=b", 
        "--nosuffix", 
        "--unquoted", 
        "-o", 
        "name,attr,cache_total_blocks,chunk_size,segtype", 
        "foo/test1"
    ], 
    "delta": "0:00:00.017432", 
    "end": "2024-12-14 18:00:48.046629", 
    "rc": 0, 
    "start": "2024-12-14 18:00:48.029197"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Saturday 14 December 2024  18:00:48 -0500 (0:00:00.628)       0:06:35.681 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    }, 
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Saturday 14 December 2024  18:00:48 -0500 (0:00:00.087)       0:06:35.768 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Saturday 14 December 2024  18:00:48 -0500 (0:00:00.074)       0:06:35.842 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Saturday 14 December 2024  18:00:48 -0500 (0:00:00.063)       0:06:35.906 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Saturday 14 December 2024  18:00:48 -0500 (0:00:00.054)       0:06:35.961 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Saturday 14 December 2024  18:00:48 -0500 (0:00:00.050)       0:06:36.011 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Saturday 14 December 2024  18:00:48 -0500 (0:00:00.065)       0:06:36.077 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    }, 
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Saturday 14 December 2024  18:00:48 -0500 (0:00:00.065)       0:06:36.143 ***** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Saturday 14 December 2024  18:00:48 -0500 (0:00:00.062)       0:06:36.205 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null, 
        "storage_test_crypttab": null, 
        "storage_test_fstab": null
    }, 
    "changed": false
}

TASK [Create a file] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml:12
Saturday 14 December 2024  18:00:48 -0500 (0:00:00.056)       0:06:36.262 ***** 
changed: [managed-node2] => {
    "changed": true, 
    "dest": "/opt/test1/quux", 
    "gid": 0, 
    "group": "root", 
    "mode": "0644", 
    "owner": "root", 
    "secontext": "unconfined_u:object_r:unlabeled_t:s0", 
    "size": 0, 
    "state": "file", 
    "uid": 0
}

TASK [Test for correct handling of safe_mode] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:423
Saturday 14 December 2024  18:00:49 -0500 (0:00:00.434)       0:06:36.697 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node2

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Saturday 14 December 2024  18:00:49 -0500 (0:00:00.167)       0:06:36.864 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_pools_global": [], 
        "storage_safe_mode_global": true, 
        "storage_volumes_global": []
    }, 
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Saturday 14 December 2024  18:00:49 -0500 (0:00:00.102)       0:06:36.967 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  18:00:49 -0500 (0:00:00.123)       0:06:37.090 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  18:00:49 -0500 (0:00:00.130)       0:06:37.221 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  18:00:49 -0500 (0:00:00.076)       0:06:37.298 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  18:00:50 -0500 (0:00:00.142)       0:06:37.441 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  18:00:50 -0500 (0:00:00.101)       0:06:37.543 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  18:00:50 -0500 (0:00:00.111)       0:06:37.654 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  18:00:50 -0500 (0:00:00.060)       0:06:37.715 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  18:00:50 -0500 (0:00:00.070)       0:06:37.786 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  18:00:50 -0500 (0:00:00.153)       0:06:37.939 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  18:00:51 -0500 (0:00:01.409)       0:06:39.349 ***** 
ok: [managed-node2] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ], 
            "name": "foo", 
            "type": "lvm", 
            "volumes": [
                {
                    "encryption": false, 
                    "encryption_password": "yabbadabbadoo", 
                    "mount_point": "/opt/test1", 
                    "name": "test1", 
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  18:00:52 -0500 (0:00:00.047)       0:06:39.397 ***** 
ok: [managed-node2] => {
    "storage_volumes": []
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  18:00:52 -0500 (0:00:00.052)       0:06:39.449 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [
        "lvm2"
    ], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  18:00:56 -0500 (0:00:04.255)       0:06:43.705 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  18:00:56 -0500 (0:00:00.128)       0:06:43.833 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  18:00:56 -0500 (0:00:00.092)       0:06:43.925 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  18:00:56 -0500 (0:00:00.069)       0:06:43.995 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  18:00:56 -0500 (0:00:00.087)       0:06:44.083 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "7:lvm2-2.02.187-6.el7_9.5.x86_64 providing lvm2 is already installed", 
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: kpartx lvm2

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  18:00:57 -0500 (0:00:00.909)       0:06:44.992 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service": {
                "name": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  18:00:58 -0500 (0:00:01.113)       0:06:46.106 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": [
            "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service"
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  18:00:58 -0500 (0:00:00.098)       0:06:46.205 ***** 
changed: [managed-node2] => (item=systemd-cryptsetup@luks\x2d3365bc93\x2d2fc0\x2d47e8\x2db784\x2db63a2126f78c.service) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
    "name": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "After": "systemd-journald.socket systemd-readahead-replay.service cryptsetup-pre.target dev-mapper-foo\\x2dtest1.device system-systemd\\x2dcryptsetup.slice systemd-readahead-collect.service", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "Before": "cryptsetup.target umount.target", 
        "BindsTo": "dev-mapper-foo\\x2dtest1.device", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "Conflicts": "umount.target", 
        "ControlPID": "0", 
        "DefaultDependencies": "no", 
        "Delegate": "no", 
        "Description": "Cryptography Setup for luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
        "DevicePolicy": "auto", 
        "Documentation": "man:crypttab(5) man:systemd-cryptsetup-generator(8) man:systemd-cryptsetup@.service(8)", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-3365bc93-2fc0-47e8-b784-b63a2126f78c /dev/mapper/foo-test1 -  ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-3365bc93-2fc0-47e8-b784-b63a2126f78c ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
        "IgnoreOnIsolate": "yes", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "4096", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "loaded", 
        "MainPID": "0", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "yes", 
        "RequiredBy": "cryptsetup.target dev-mapper-luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.device", 
        "Requires": "system-systemd\\x2dcryptsetup.slice", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "Slice": "system-systemd\\x2dcryptsetup.slice", 
        "SourcePath": "/etc/crypttab", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "journal", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "0", 
        "TimeoutStopUSec": "0", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "Type": "oneshot", 
        "UMask": "0022", 
        "UnitFilePreset": "disabled", 
        "UnitFileState": "bad", 
        "WantedBy": "dev-mapper-foo\\x2dtest1.device", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  18:00:59 -0500 (0:00:00.671)       0:06:46.876 ***** 
fatal: [managed-node2]: FAILED! => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [], 
    "pools": [], 
    "volumes": []
}

MSG:

cannot remove existing formatting on device 'luks-3365bc93-2fc0-47e8-b784-b63a2126f78c' in safe mode due to encryption removal

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Saturday 14 December 2024  18:01:03 -0500 (0:00:04.348)       0:06:51.225 ***** 
fatal: [managed-node2]: FAILED! => {
    "changed": false
}

MSG:

{u'_ansible_no_log': False, u'crypts': [], u'pools': [], u'leaves': [], u'changed': False, u'actions': [], u'failed': True, u'volumes': [], u'invocation': {u'module_args': {u'packages_only': False, u'disklabel_type': None, u'diskvolume_mkfs_option_map': {u'ext4': u'-F', u'ext3': u'-F', u'ext2': u'-F'}, u'safe_mode': True, u'pools': [{u'raid_metadata_version': None, u'encryption_key_size': None, u'encryption_key': None, u'encryption_luks_version': None, u'encryption_tang_url': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_tang_thumbprint': None, u'name': u'foo', u'encryption_password': None, u'encryption': False, u'disks': [u'sda'], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [{u'raid_metadata_version': None, u'mount_device_identifier': u'uuid', u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': u'4g', u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'encryption': False, u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'type': u'lvm', u'encryption_key_size': 0, u'deduplication': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': u'luks1', u'raid_stripe_size': None, u'mount_passno': 0, u'mount_user': None, u'raid_spare_count': None, u'name': u'test1', u'cache_mode': None, u'raid_disks': [], u'mount_group': None, u'fs_overwrite_existing': True, u'disks': [u'sda'], u'cached': False, u'thin_pool_size': None, u'thin': False, u'mount_check': 0, u'cache_size': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}], u'shared': False, u'encryption_clevis_pin': None, u'type': u'lvm', u'encryption_cipher': None, u'raid_chunk_size': None}], u'volumes': [], u'pool_defaults': {u'raid_metadata_version': None, u'encryption_cipher': None, u'encryption_key': None, u'encryption_luks_version': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_password': None, u'encryption': False, u'disks': [], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [], u'shared': False, u'type': u'lvm', u'encryption_key_size': None, u'raid_chunk_size': None}, u'volume_defaults': {u'raid_metadata_version': None, u'raid_level': None, u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': 0, u'mount_point': u'', u'compression': None, u'encryption_password': None, u'encryption': False, u'mount_device_identifier': u'uuid', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'thin_pool_name': None, u'type': u'lvm', u'encryption_key_size': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'cache_size': 0, u'raid_spare_count': None, u'cache_mode': None, u'deduplication': None, u'cached': False, u'fs_overwrite_existing': True, u'disks': [], u'thin_pool_size': None, u'thin': None, u'mount_check': 0, u'mount_passno': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}, u'use_partitions': None}}, u'mounts': [], u'packages': [], u'msg': u"cannot remove existing formatting on device 'luks-3365bc93-2fc0-47e8-b784-b63a2126f78c' in safe mode due to encryption removal"}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  18:01:04 -0500 (0:00:00.327)       0:06:51.552 ***** 
changed: [managed-node2] => (item=systemd-cryptsetup@luks\x2d3365bc93\x2d2fc0\x2d47e8\x2db784\x2db63a2126f78c.service) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
    "name": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "ControlPID": "0", 
        "DefaultDependencies": "yes", 
        "Delegate": "no", 
        "Description": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
        "DevicePolicy": "auto", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/dev/null", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
        "IgnoreOnIsolate": "no", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "1048576", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "masked", 
        "MainPID": "0", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "no", 
        "RequiredBy": "dev-mapper-luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.device cryptsetup.target", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "inherit", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "1min 30s", 
        "TimeoutStopUSec": "1min 30s", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "UMask": "0022", 
        "UnitFileState": "bad", 
        "WantedBy": "dev-mapper-foo\\x2dtest1.device", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Saturday 14 December 2024  18:01:04 -0500 (0:00:00.718)       0:06:52.271 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Saturday 14 December 2024  18:01:04 -0500 (0:00:00.065)       0:06:52.337 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Saturday 14 December 2024  18:01:05 -0500 (0:00:00.088)       0:06:52.425 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Stat the file] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:11
Saturday 14 December 2024  18:01:05 -0500 (0:00:00.052)       0:06:52.477 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217249.2167041, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", 
        "ctime": 1734217249.2167041, 
        "dev": 64769, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 67, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/x-empty", 
        "mode": "0644", 
        "mtime": 1734217249.2167041, 
        "nlink": 1, 
        "path": "/opt/test1/quux", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": "144112914", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Assert file presence] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:16
Saturday 14 December 2024  18:01:05 -0500 (0:00:00.541)       0:06:53.019 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Remove the encryption layer] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:446
Saturday 14 December 2024  18:01:05 -0500 (0:00:00.078)       0:06:53.098 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  18:01:05 -0500 (0:00:00.167)       0:06:53.265 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  18:01:05 -0500 (0:00:00.092)       0:06:53.358 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  18:01:06 -0500 (0:00:00.066)       0:06:53.425 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  18:01:06 -0500 (0:00:00.157)       0:06:53.582 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  18:01:06 -0500 (0:00:00.055)       0:06:53.637 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  18:01:06 -0500 (0:00:00.055)       0:06:53.692 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  18:01:06 -0500 (0:00:00.057)       0:06:53.750 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  18:01:06 -0500 (0:00:00.060)       0:06:53.810 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  18:01:06 -0500 (0:00:00.132)       0:06:53.943 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  18:01:08 -0500 (0:00:01.681)       0:06:55.624 ***** 
ok: [managed-node2] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ], 
            "name": "foo", 
            "type": "lvm", 
            "volumes": [
                {
                    "encryption": false, 
                    "encryption_password": "yabbadabbadoo", 
                    "mount_point": "/opt/test1", 
                    "name": "test1", 
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  18:01:08 -0500 (0:00:00.078)       0:06:55.703 ***** 
ok: [managed-node2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  18:01:08 -0500 (0:00:00.120)       0:06:55.823 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [
        "lvm2"
    ], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  18:01:12 -0500 (0:00:04.490)       0:07:00.314 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  18:01:13 -0500 (0:00:00.184)       0:07:00.498 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  18:01:13 -0500 (0:00:00.114)       0:07:00.613 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  18:01:13 -0500 (0:00:00.087)       0:07:00.701 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  18:01:13 -0500 (0:00:00.090)       0:07:00.791 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "7:lvm2-2.02.187-6.el7_9.5.x86_64 providing lvm2 is already installed", 
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: kpartx lvm2

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  18:01:14 -0500 (0:00:00.790)       0:07:01.581 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service": {
                "name": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  18:01:15 -0500 (0:00:01.370)       0:07:02.952 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": [
            "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service"
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  18:01:15 -0500 (0:00:00.316)       0:07:03.269 ***** 
changed: [managed-node2] => (item=systemd-cryptsetup@luks\x2d3365bc93\x2d2fc0\x2d47e8\x2db784\x2db63a2126f78c.service) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
    "name": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "After": "system-systemd\\x2dcryptsetup.slice systemd-readahead-replay.service systemd-journald.socket systemd-readahead-collect.service cryptsetup-pre.target dev-mapper-foo\\x2dtest1.device", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "Before": "cryptsetup.target umount.target", 
        "BindsTo": "dev-mapper-foo\\x2dtest1.device", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "Conflicts": "umount.target", 
        "ControlPID": "0", 
        "DefaultDependencies": "no", 
        "Delegate": "no", 
        "Description": "Cryptography Setup for luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
        "DevicePolicy": "auto", 
        "Documentation": "man:crypttab(5) man:systemd-cryptsetup-generator(8) man:systemd-cryptsetup@.service(8)", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-3365bc93-2fc0-47e8-b784-b63a2126f78c /dev/mapper/foo-test1 -  ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-3365bc93-2fc0-47e8-b784-b63a2126f78c ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
        "IgnoreOnIsolate": "yes", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "4096", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "loaded", 
        "MainPID": "0", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "yes", 
        "RequiredBy": "dev-mapper-luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.device cryptsetup.target", 
        "Requires": "system-systemd\\x2dcryptsetup.slice", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "Slice": "system-systemd\\x2dcryptsetup.slice", 
        "SourcePath": "/etc/crypttab", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "journal", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "0", 
        "TimeoutStopUSec": "0", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "Type": "oneshot", 
        "UMask": "0022", 
        "UnitFilePreset": "disabled", 
        "UnitFileState": "bad", 
        "WantedBy": "dev-mapper-foo\\x2dtest1.device", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  18:01:16 -0500 (0:00:00.905)       0:07:04.175 ***** 
changed: [managed-node2] => {
    "actions": [
        {
            "action": "destroy format", 
            "device": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
            "fs_type": "xfs"
        }, 
        {
            "action": "destroy device", 
            "device": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
            "fs_type": null
        }, 
        {
            "action": "destroy format", 
            "device": "/dev/mapper/foo-test1", 
            "fs_type": "luks"
        }, 
        {
            "action": "create format", 
            "device": "/dev/mapper/foo-test1", 
            "fs_type": "xfs"
        }
    ], 
    "changed": true, 
    "crypts": [
        {
            "backing_device": "/dev/mapper/foo-test1", 
            "name": "luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
            "password": "-", 
            "state": "absent"
        }
    ], 
    "leaves": [
        "/dev/mapper/foo-test1", 
        "/dev/sdb", 
        "/dev/sdc", 
        "/dev/sdd", 
        "/dev/sde", 
        "/dev/sdf", 
        "/dev/sdg", 
        "/dev/sdh", 
        "/dev/sdi", 
        "/dev/xvda1"
    ], 
    "mounts": [
        {
            "fstype": "xfs", 
            "path": "/opt/test1", 
            "src": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
            "state": "absent"
        }, 
        {
            "dump": 0, 
            "fstype": "xfs", 
            "group": null, 
            "mode": null, 
            "opts": "defaults", 
            "owner": null, 
            "passno": 0, 
            "path": "/opt/test1", 
            "src": "/dev/mapper/foo-test1", 
            "state": "mounted"
        }
    ], 
    "packages": [
        "xfsprogs", 
        "e2fsprogs", 
        "lvm2"
    ], 
    "pools": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_cipher": null, 
            "encryption_clevis_pin": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": null, 
            "encryption_tang_thumbprint": null, 
            "encryption_tang_url": null, 
            "grow_to_fill": false, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "shared": false, 
            "state": "present", 
            "type": "lvm", 
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1", 
                    "_kernel_device": "/dev/dm-0", 
                    "_mount_id": "/dev/mapper/foo-test1", 
                    "_raw_device": "/dev/mapper/foo-test1", 
                    "_raw_kernel_device": "/dev/dm-0", 
                    "cache_devices": [], 
                    "cache_mode": null, 
                    "cache_size": 0, 
                    "cached": false, 
                    "compression": null, 
                    "deduplication": null, 
                    "disks": [
                        "sda"
                    ], 
                    "encryption": false, 
                    "encryption_cipher": null, 
                    "encryption_key": null, 
                    "encryption_key_size": 0, 
                    "encryption_luks_version": "luks1", 
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                    "fs_create_options": "", 
                    "fs_label": "", 
                    "fs_overwrite_existing": true, 
                    "fs_type": "xfs", 
                    "mount_check": 0, 
                    "mount_device_identifier": "uuid", 
                    "mount_group": null, 
                    "mount_mode": null, 
                    "mount_options": "defaults", 
                    "mount_passno": 0, 
                    "mount_point": "/opt/test1", 
                    "mount_user": null, 
                    "name": "test1", 
                    "raid_chunk_size": null, 
                    "raid_device_count": null, 
                    "raid_disks": [], 
                    "raid_level": null, 
                    "raid_metadata_version": null, 
                    "raid_spare_count": null, 
                    "raid_stripe_size": null, 
                    "size": "4g", 
                    "state": "present", 
                    "thin": false, 
                    "thin_pool_name": null, 
                    "thin_pool_size": null, 
                    "type": "lvm", 
                    "vdo_pool_size": null
                }
            ]
        }
    ], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Saturday 14 December 2024  18:02:22 -0500 (0:01:05.269)       0:08:09.444 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Saturday 14 December 2024  18:02:22 -0500 (0:00:00.056)       0:08:09.501 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217188.1247318, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "7e4e8b3ad4d6e63ef5f280d635ca488627b2dd22", 
        "ctime": 1734217188.1217318, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 263588, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0644", 
        "mtime": 1734217188.1217318, 
        "nlink": 1, 
        "path": "/etc/fstab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 1311, 
        "uid": 0, 
        "version": "606402765", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Saturday 14 December 2024  18:02:22 -0500 (0:00:00.394)       0:08:09.896 ***** 
ok: [managed-node2] => {
    "backup": "", 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  18:02:22 -0500 (0:00:00.381)       0:08:10.277 ***** 
changed: [managed-node2] => (item=systemd-cryptsetup@luks\x2d3365bc93\x2d2fc0\x2d47e8\x2db784\x2db63a2126f78c.service) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
    "name": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "ControlPID": "0", 
        "DefaultDependencies": "yes", 
        "Delegate": "no", 
        "Description": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
        "DevicePolicy": "auto", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/dev/null", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
        "IgnoreOnIsolate": "no", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "1048576", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "masked", 
        "MainPID": "0", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "no", 
        "RequiredBy": "cryptsetup.target dev-mapper-luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.device", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "inherit", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "1min 30s", 
        "TimeoutStopUSec": "1min 30s", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "UMask": "0022", 
        "UnitFileState": "bad", 
        "WantedBy": "dev-mapper-foo\\x2dtest1.device", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Saturday 14 December 2024  18:02:23 -0500 (0:00:00.504)       0:08:10.782 ***** 
ok: [managed-node2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format", 
                "device": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                "fs_type": "xfs"
            }, 
            {
                "action": "destroy device", 
                "device": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                "fs_type": null
            }, 
            {
                "action": "destroy format", 
                "device": "/dev/mapper/foo-test1", 
                "fs_type": "luks"
            }, 
            {
                "action": "create format", 
                "device": "/dev/mapper/foo-test1", 
                "fs_type": "xfs"
            }
        ], 
        "changed": true, 
        "crypts": [
            {
                "backing_device": "/dev/mapper/foo-test1", 
                "name": "luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                "password": "-", 
                "state": "absent"
            }
        ], 
        "failed": false, 
        "leaves": [
            "/dev/mapper/foo-test1", 
            "/dev/sdb", 
            "/dev/sdc", 
            "/dev/sdd", 
            "/dev/sde", 
            "/dev/sdf", 
            "/dev/sdg", 
            "/dev/sdh", 
            "/dev/sdi", 
            "/dev/xvda1"
        ], 
        "mounts": [
            {
                "fstype": "xfs", 
                "path": "/opt/test1", 
                "src": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
                "state": "absent"
            }, 
            {
                "dump": 0, 
                "fstype": "xfs", 
                "group": null, 
                "mode": null, 
                "opts": "defaults", 
                "owner": null, 
                "passno": 0, 
                "path": "/opt/test1", 
                "src": "/dev/mapper/foo-test1", 
                "state": "mounted"
            }
        ], 
        "packages": [
            "xfsprogs", 
            "e2fsprogs", 
            "lvm2"
        ], 
        "pools": [
            {
                "disks": [
                    "sda"
                ], 
                "encryption": false, 
                "encryption_cipher": null, 
                "encryption_clevis_pin": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": null, 
                "encryption_tang_thumbprint": null, 
                "encryption_tang_url": null, 
                "grow_to_fill": false, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "shared": false, 
                "state": "present", 
                "type": "lvm", 
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1", 
                        "_kernel_device": "/dev/dm-0", 
                        "_mount_id": "/dev/mapper/foo-test1", 
                        "_raw_device": "/dev/mapper/foo-test1", 
                        "_raw_kernel_device": "/dev/dm-0", 
                        "cache_devices": [], 
                        "cache_mode": null, 
                        "cache_size": 0, 
                        "cached": false, 
                        "compression": null, 
                        "deduplication": null, 
                        "disks": [
                            "sda"
                        ], 
                        "encryption": false, 
                        "encryption_cipher": null, 
                        "encryption_key": null, 
                        "encryption_key_size": 0, 
                        "encryption_luks_version": "luks1", 
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                        "fs_create_options": "", 
                        "fs_label": "", 
                        "fs_overwrite_existing": true, 
                        "fs_type": "xfs", 
                        "mount_check": 0, 
                        "mount_device_identifier": "uuid", 
                        "mount_group": null, 
                        "mount_mode": null, 
                        "mount_options": "defaults", 
                        "mount_passno": 0, 
                        "mount_point": "/opt/test1", 
                        "mount_user": null, 
                        "name": "test1", 
                        "raid_chunk_size": null, 
                        "raid_device_count": null, 
                        "raid_disks": [], 
                        "raid_level": null, 
                        "raid_metadata_version": null, 
                        "raid_spare_count": null, 
                        "raid_stripe_size": null, 
                        "size": "4g", 
                        "state": "present", 
                        "thin": false, 
                        "thin_pool_name": null, 
                        "thin_pool_size": null, 
                        "type": "lvm", 
                        "vdo_pool_size": null
                    }
                ]
            }
        ], 
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Saturday 14 December 2024  18:02:23 -0500 (0:00:00.051)       0:08:10.833 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ], 
                "encryption": false, 
                "encryption_cipher": null, 
                "encryption_clevis_pin": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": null, 
                "encryption_tang_thumbprint": null, 
                "encryption_tang_url": null, 
                "grow_to_fill": false, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "shared": false, 
                "state": "present", 
                "type": "lvm", 
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1", 
                        "_kernel_device": "/dev/dm-0", 
                        "_mount_id": "/dev/mapper/foo-test1", 
                        "_raw_device": "/dev/mapper/foo-test1", 
                        "_raw_kernel_device": "/dev/dm-0", 
                        "cache_devices": [], 
                        "cache_mode": null, 
                        "cache_size": 0, 
                        "cached": false, 
                        "compression": null, 
                        "deduplication": null, 
                        "disks": [
                            "sda"
                        ], 
                        "encryption": false, 
                        "encryption_cipher": null, 
                        "encryption_key": null, 
                        "encryption_key_size": 0, 
                        "encryption_luks_version": "luks1", 
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                        "fs_create_options": "", 
                        "fs_label": "", 
                        "fs_overwrite_existing": true, 
                        "fs_type": "xfs", 
                        "mount_check": 0, 
                        "mount_device_identifier": "uuid", 
                        "mount_group": null, 
                        "mount_mode": null, 
                        "mount_options": "defaults", 
                        "mount_passno": 0, 
                        "mount_point": "/opt/test1", 
                        "mount_user": null, 
                        "name": "test1", 
                        "raid_chunk_size": null, 
                        "raid_device_count": null, 
                        "raid_disks": [], 
                        "raid_level": null, 
                        "raid_metadata_version": null, 
                        "raid_spare_count": null, 
                        "raid_stripe_size": null, 
                        "size": "4g", 
                        "state": "present", 
                        "thin": false, 
                        "thin_pool_name": null, 
                        "thin_pool_size": null, 
                        "type": "lvm", 
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Saturday 14 December 2024  18:02:23 -0500 (0:00:00.048)       0:08:10.882 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Saturday 14 December 2024  18:02:23 -0500 (0:00:00.043)       0:08:10.926 ***** 
changed: [managed-node2] => (item={u'src': u'/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c', u'state': u'absent', u'fstype': u'xfs', u'path': u'/opt/test1'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": true, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "fstype": "xfs", 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
        "state": "absent"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "/dev/mapper/luks-3365bc93-2fc0-47e8-b784-b63a2126f78c"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Saturday 14 December 2024  18:02:23 -0500 (0:00:00.423)       0:08:11.349 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Saturday 14 December 2024  18:02:24 -0500 (0:00:00.622)       0:08:11.972 ***** 
changed: [managed-node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": true, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "/dev/mapper/foo-test1", 
        "state": "mounted"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Saturday 14 December 2024  18:02:25 -0500 (0:00:00.539)       0:08:12.511 ***** 
skipping: [managed-node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'})  => {
    "ansible_loop_var": "mount_info", 
    "changed": false, 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "/dev/mapper/foo-test1", 
        "state": "mounted"
    }, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Saturday 14 December 2024  18:02:25 -0500 (0:00:00.087)       0:08:12.599 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Saturday 14 December 2024  18:02:25 -0500 (0:00:00.507)       0:08:13.107 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217191.81573, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "184e4520d6a02c13cdc2844673950a75c9ab85fd", 
        "ctime": 1734217189.7647312, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 917511, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0600", 
        "mtime": 1734217189.7647312, 
        "nlink": 1, 
        "path": "/etc/crypttab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": false, 
        "roth": false, 
        "rusr": true, 
        "size": 66, 
        "uid": 0, 
        "version": "606409565", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Saturday 14 December 2024  18:02:26 -0500 (0:00:00.412)       0:08:13.519 ***** 
changed: [managed-node2] => (item={u'state': u'absent', u'password': u'-', u'name': u'luks-3365bc93-2fc0-47e8-b784-b63a2126f78c', u'backing_device': u'/dev/mapper/foo-test1'}) => {
    "ansible_loop_var": "entry", 
    "backup": "", 
    "changed": true, 
    "entry": {
        "backing_device": "/dev/mapper/foo-test1", 
        "name": "luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
        "password": "-", 
        "state": "absent"
    }, 
    "found": 1
}

MSG:

1 line(s) removed

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Saturday 14 December 2024  18:02:26 -0500 (0:00:00.364)       0:08:13.884 ***** 
ok: [managed-node2]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:462
Saturday 14 December 2024  18:02:27 -0500 (0:00:00.739)       0:08:14.623 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node2

TASK [Print out pool information] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Saturday 14 December 2024  18:02:27 -0500 (0:00:00.078)       0:08:14.701 ***** 
ok: [managed-node2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_cipher": null, 
            "encryption_clevis_pin": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": null, 
            "encryption_tang_thumbprint": null, 
            "encryption_tang_url": null, 
            "grow_to_fill": false, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "shared": false, 
            "state": "present", 
            "type": "lvm", 
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1", 
                    "_kernel_device": "/dev/dm-0", 
                    "_mount_id": "/dev/mapper/foo-test1", 
                    "_raw_device": "/dev/mapper/foo-test1", 
                    "_raw_kernel_device": "/dev/dm-0", 
                    "cache_devices": [], 
                    "cache_mode": null, 
                    "cache_size": 0, 
                    "cached": false, 
                    "compression": null, 
                    "deduplication": null, 
                    "disks": [
                        "sda"
                    ], 
                    "encryption": false, 
                    "encryption_cipher": null, 
                    "encryption_key": null, 
                    "encryption_key_size": 0, 
                    "encryption_luks_version": "luks1", 
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                    "fs_create_options": "", 
                    "fs_label": "", 
                    "fs_overwrite_existing": true, 
                    "fs_type": "xfs", 
                    "mount_check": 0, 
                    "mount_device_identifier": "uuid", 
                    "mount_group": null, 
                    "mount_mode": null, 
                    "mount_options": "defaults", 
                    "mount_passno": 0, 
                    "mount_point": "/opt/test1", 
                    "mount_user": null, 
                    "name": "test1", 
                    "raid_chunk_size": null, 
                    "raid_device_count": null, 
                    "raid_disks": [], 
                    "raid_level": null, 
                    "raid_metadata_version": null, 
                    "raid_spare_count": null, 
                    "raid_stripe_size": null, 
                    "size": "4g", 
                    "state": "present", 
                    "thin": false, 
                    "thin_pool_name": null, 
                    "thin_pool_size": null, 
                    "type": "lvm", 
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Saturday 14 December 2024  18:02:27 -0500 (0:00:00.049)       0:08:14.751 ***** 
skipping: [managed-node2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Saturday 14 December 2024  18:02:27 -0500 (0:00:00.044)       0:08:14.796 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "xfs", 
            "label": "", 
            "mountpoint": "/opt/test1", 
            "name": "/dev/mapper/foo-test1", 
            "size": "4G", 
            "type": "lvm", 
            "uuid": "ec70cb4b-d6d1-4ae3-9d4b-f758b8c8a3ac"
        }, 
        "/dev/sda": {
            "fstype": "LVM2_member", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sda", 
            "size": "10G", 
            "type": "disk", 
            "uuid": "xa7q16-X4RQ-vFWY-Ftcj-AFkA-0I7B-z0LDg5"
        }, 
        "/dev/sdb": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdb", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdc": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdc", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdd": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdd", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sde": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sde", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdf": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdf", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdg": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdg", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdh": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdh", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdi": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdi", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/xvda", 
            "size": "250G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda1": {
            "fstype": "ext4", 
            "label": "", 
            "mountpoint": "/", 
            "name": "/dev/xvda1", 
            "size": "250G", 
            "type": "partition", 
            "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Saturday 14 December 2024  18:02:27 -0500 (0:00:00.391)       0:08:15.187 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/fstab"
    ], 
    "delta": "0:00:00.002920", 
    "end": "2024-12-14 18:02:28.217912", 
    "rc": 0, 
    "start": "2024-12-14 18:02:28.214992"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Thu Jun 20 10:23:46 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 /                       ext4    defaults        1 1
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Saturday 14 December 2024  18:02:28 -0500 (0:00:00.512)       0:08:15.699 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/crypttab"
    ], 
    "delta": "0:00:00.002818", 
    "end": "2024-12-14 18:02:28.642083", 
    "failed_when_result": false, 
    "rc": 0, 
    "start": "2024-12-14 18:02:28.639265"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Saturday 14 December 2024  18:02:28 -0500 (0:00:00.409)       0:08:16.109 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node2

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Saturday 14 December 2024  18:02:28 -0500 (0:00:00.084)       0:08:16.193 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members", 
            "volumes"
        ]
    }, 
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Saturday 14 December 2024  18:02:28 -0500 (0:00:00.048)       0:08:16.242 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "vgs", 
        "--noheadings", 
        "--binary", 
        "-o", 
        "shared", 
        "foo"
    ], 
    "delta": "0:00:00.017025", 
    "end": "2024-12-14 18:02:29.294579", 
    "rc": 0, 
    "start": "2024-12-14 18:02:29.277554"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Saturday 14 December 2024  18:02:29 -0500 (0:00:00.523)       0:08:16.765 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Saturday 14 December 2024  18:02:29 -0500 (0:00:00.081)       0:08:16.847 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node2

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Saturday 14 December 2024  18:02:29 -0500 (0:00:00.130)       0:08:16.977 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1", 
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    }, 
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Saturday 14 December 2024  18:02:29 -0500 (0:00:00.096)       0:08:17.074 ***** 
ok: [managed-node2] => (item=/dev/sda) => {
    "ansible_loop_var": "pv", 
    "changed": false, 
    "device": "/dev/sda", 
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Saturday 14 December 2024  18:02:30 -0500 (0:00:00.641)       0:08:17.715 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    }, 
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Saturday 14 December 2024  18:02:30 -0500 (0:00:00.080)       0:08:17.796 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    }, 
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Saturday 14 December 2024  18:02:30 -0500 (0:00:00.083)       0:08:17.879 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Saturday 14 December 2024  18:02:30 -0500 (0:00:00.075)       0:08:17.954 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    }, 
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Saturday 14 December 2024  18:02:30 -0500 (0:00:00.068)       0:08:18.023 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    }, 
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Saturday 14 December 2024  18:02:30 -0500 (0:00:00.068)       0:08:18.092 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Saturday 14 December 2024  18:02:30 -0500 (0:00:00.059)       0:08:18.151 ***** 
ok: [managed-node2] => (item=/dev/sda) => {
    "ansible_loop_var": "pv", 
    "changed": false, 
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Saturday 14 December 2024  18:02:30 -0500 (0:00:00.079)       0:08:18.231 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.8.171 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Saturday 14 December 2024  18:02:31 -0500 (0:00:00.373)       0:08:18.604 ***** 
skipping: [managed-node2] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv", 
    "changed": false, 
    "skip_reason": "Conditional result was False", 
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Saturday 14 December 2024  18:02:31 -0500 (0:00:00.070)       0:08:18.675 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node2

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Saturday 14 December 2024  18:02:31 -0500 (0:00:00.110)       0:08:18.785 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Saturday 14 December 2024  18:02:31 -0500 (0:00:00.039)       0:08:18.825 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Saturday 14 December 2024  18:02:31 -0500 (0:00:00.045)       0:08:18.871 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Saturday 14 December 2024  18:02:31 -0500 (0:00:00.067)       0:08:18.938 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Saturday 14 December 2024  18:02:31 -0500 (0:00:00.082)       0:08:19.021 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Saturday 14 December 2024  18:02:31 -0500 (0:00:00.088)       0:08:19.110 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Saturday 14 December 2024  18:02:31 -0500 (0:00:00.107)       0:08:19.217 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Saturday 14 December 2024  18:02:31 -0500 (0:00:00.091)       0:08:19.309 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Saturday 14 December 2024  18:02:31 -0500 (0:00:00.060)       0:08:19.370 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Saturday 14 December 2024  18:02:32 -0500 (0:00:00.068)       0:08:19.438 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Saturday 14 December 2024  18:02:32 -0500 (0:00:00.054)       0:08:19.493 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null, 
        "storage_test_md_chunk_size_re": null, 
        "storage_test_md_metadata_version_re": null, 
        "storage_test_md_spare_devices_re": null
    }, 
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Saturday 14 December 2024  18:02:32 -0500 (0:00:00.070)       0:08:19.571 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Saturday 14 December 2024  18:02:32 -0500 (0:00:00.291)       0:08:19.863 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node2

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Saturday 14 December 2024  18:02:32 -0500 (0:00:00.160)       0:08:20.023 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Saturday 14 December 2024  18:02:32 -0500 (0:00:00.066)       0:08:20.090 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Saturday 14 December 2024  18:02:32 -0500 (0:00:00.069)       0:08:20.159 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Saturday 14 December 2024  18:02:32 -0500 (0:00:00.078)       0:08:20.238 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Saturday 14 December 2024  18:02:32 -0500 (0:00:00.071)       0:08:20.310 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Saturday 14 December 2024  18:02:32 -0500 (0:00:00.060)       0:08:20.371 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Saturday 14 December 2024  18:02:33 -0500 (0:00:00.062)       0:08:20.433 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Saturday 14 December 2024  18:02:33 -0500 (0:00:00.058)       0:08:20.491 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Saturday 14 December 2024  18:02:33 -0500 (0:00:00.136)       0:08:20.628 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node2

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Saturday 14 December 2024  18:02:33 -0500 (0:00:00.159)       0:08:20.787 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Saturday 14 December 2024  18:02:33 -0500 (0:00:00.068)       0:08:20.856 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Saturday 14 December 2024  18:02:33 -0500 (0:00:00.068)       0:08:20.924 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Saturday 14 December 2024  18:02:33 -0500 (0:00:00.083)       0:08:21.008 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    }, 
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Saturday 14 December 2024  18:02:33 -0500 (0:00:00.057)       0:08:21.066 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node2

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Saturday 14 December 2024  18:02:33 -0500 (0:00:00.129)       0:08:21.195 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0", 
        "_storage_test_expected_crypttab_key_file": "-"
    }, 
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Saturday 14 December 2024  18:02:33 -0500 (0:00:00.065)       0:08:21.261 ***** 
skipping: [managed-node2] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda", 
    "ansible_loop_var": "_storage_test_pool_member_path", 
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Saturday 14 December 2024  18:02:33 -0500 (0:00:00.068)       0:08:21.330 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node2

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Saturday 14 December 2024  18:02:34 -0500 (0:00:00.109)       0:08:21.439 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    }, 
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Saturday 14 December 2024  18:02:34 -0500 (0:00:00.062)       0:08:21.501 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Saturday 14 December 2024  18:02:34 -0500 (0:00:00.063)       0:08:21.565 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Saturday 14 December 2024  18:02:34 -0500 (0:00:00.058)       0:08:21.624 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Saturday 14 December 2024  18:02:34 -0500 (0:00:00.135)       0:08:21.759 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Saturday 14 December 2024  18:02:34 -0500 (0:00:00.055)       0:08:21.815 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    }, 
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Saturday 14 December 2024  18:02:34 -0500 (0:00:00.054)       0:08:21.870 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null, 
        "_storage_test_crypttab_key_file": null
    }, 
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Saturday 14 December 2024  18:02:34 -0500 (0:00:00.054)       0:08:21.924 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Saturday 14 December 2024  18:02:34 -0500 (0:00:00.152)       0:08:22.077 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node2

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Saturday 14 December 2024  18:02:34 -0500 (0:00:00.122)       0:08:22.200 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Saturday 14 December 2024  18:02:34 -0500 (0:00:00.056)       0:08:22.256 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Saturday 14 December 2024  18:02:34 -0500 (0:00:00.056)       0:08:22.312 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Saturday 14 December 2024  18:02:34 -0500 (0:00:00.062)       0:08:22.375 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Saturday 14 December 2024  18:02:35 -0500 (0:00:00.057)       0:08:22.432 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Saturday 14 December 2024  18:02:35 -0500 (0:00:00.058)       0:08:22.490 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Saturday 14 December 2024  18:02:35 -0500 (0:00:00.052)       0:08:22.543 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    }, 
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Saturday 14 December 2024  18:02:35 -0500 (0:00:00.055)       0:08:22.599 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node2

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Saturday 14 December 2024  18:02:35 -0500 (0:00:00.145)       0:08:22.744 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Saturday 14 December 2024  18:02:35 -0500 (0:00:00.056)       0:08:22.800 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Saturday 14 December 2024  18:02:35 -0500 (0:00:00.057)       0:08:22.858 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Saturday 14 December 2024  18:02:35 -0500 (0:00:00.062)       0:08:22.921 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Saturday 14 December 2024  18:02:35 -0500 (0:00:00.052)       0:08:22.973 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Saturday 14 December 2024  18:02:35 -0500 (0:00:00.058)       0:08:23.031 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    }, 
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Saturday 14 December 2024  18:02:35 -0500 (0:00:00.093)       0:08:23.125 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null, 
        "_storage_test_expected_pv_count": null, 
        "_storage_test_expected_pv_type": null, 
        "_storage_test_pool_pvs": [], 
        "_storage_test_pool_pvs_lvm": []
    }, 
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Saturday 14 December 2024  18:02:35 -0500 (0:00:00.063)       0:08:23.188 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node2

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Saturday 14 December 2024  18:02:35 -0500 (0:00:00.116)       0:08:23.305 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true, 
        "_storage_volume_tests": [
            "mount", 
            "fstab", 
            "fs", 
            "device", 
            "encryption", 
            "md", 
            "size", 
            "cache"
        ]
    }, 
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Saturday 14 December 2024  18:02:35 -0500 (0:00:00.073)       0:08:23.378 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node2

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Saturday 14 December 2024  18:02:36 -0500 (0:00:00.267)       0:08:23.646 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    }, 
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Saturday 14 December 2024  18:02:36 -0500 (0:00:00.065)       0:08:23.712 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1", 
        "storage_test_swap_expected_matches": "0"
    }, 
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Saturday 14 December 2024  18:02:36 -0500 (0:00:00.069)       0:08:23.781 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Saturday 14 December 2024  18:02:36 -0500 (0:00:00.057)       0:08:23.838 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Saturday 14 December 2024  18:02:36 -0500 (0:00:00.070)       0:08:23.909 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Saturday 14 December 2024  18:02:36 -0500 (0:00:00.064)       0:08:23.974 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Saturday 14 December 2024  18:02:36 -0500 (0:00:00.059)       0:08:24.034 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Saturday 14 December 2024  18:02:36 -0500 (0:00:00.059)       0:08:24.093 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Saturday 14 December 2024  18:02:36 -0500 (0:00:00.074)       0:08:24.168 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Saturday 14 December 2024  18:02:36 -0500 (0:00:00.055)       0:08:24.224 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Saturday 14 December 2024  18:02:36 -0500 (0:00:00.054)       0:08:24.278 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null, 
        "storage_test_mount_expected_mount_point": null, 
        "storage_test_swap_expected_matches": null, 
        "storage_test_swaps": null, 
        "storage_test_sys_node": null
    }, 
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Saturday 14 December 2024  18:02:36 -0500 (0:00:00.062)       0:08:24.341 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1", 
        "storage_test_fstab_expected_mount_options_matches": "1", 
        "storage_test_fstab_expected_mount_point_matches": "1", 
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ], 
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ], 
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    }, 
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Saturday 14 December 2024  18:02:37 -0500 (0:00:00.065)       0:08:24.406 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Saturday 14 December 2024  18:02:37 -0500 (0:00:00.047)       0:08:24.454 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Saturday 14 December 2024  18:02:37 -0500 (0:00:00.043)       0:08:24.497 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Saturday 14 December 2024  18:02:37 -0500 (0:00:00.035)       0:08:24.533 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Saturday 14 December 2024  18:02:37 -0500 (0:00:00.046)       0:08:24.579 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null, 
        "storage_test_fstab_expected_mount_options_matches": null, 
        "storage_test_fstab_expected_mount_point_matches": null, 
        "storage_test_fstab_id_matches": null, 
        "storage_test_fstab_mount_options_matches": null, 
        "storage_test_fstab_mount_point_matches": null
    }, 
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Saturday 14 December 2024  18:02:37 -0500 (0:00:00.056)       0:08:24.635 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Saturday 14 December 2024  18:02:37 -0500 (0:00:00.054)       0:08:24.690 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Saturday 14 December 2024  18:02:37 -0500 (0:00:00.054)       0:08:24.744 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217341.9356625, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "ctime": 1734217341.9356625, 
        "dev": 5, 
        "device_type": 64768, 
        "executable": false, 
        "exists": true, 
        "gid": 6, 
        "gr_name": "disk", 
        "inode": 453486, 
        "isblk": true, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": false, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/symlink", 
        "mode": "0660", 
        "mtime": 1734217341.9356625, 
        "nlink": 1, 
        "path": "/dev/mapper/foo-test1", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": null, 
        "wgrp": true, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Saturday 14 December 2024  18:02:37 -0500 (0:00:00.328)       0:08:25.073 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Saturday 14 December 2024  18:02:37 -0500 (0:00:00.046)       0:08:25.119 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Saturday 14 December 2024  18:02:37 -0500 (0:00:00.037)       0:08:25.157 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Saturday 14 December 2024  18:02:37 -0500 (0:00:00.063)       0:08:25.220 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    }, 
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Saturday 14 December 2024  18:02:37 -0500 (0:00:00.061)       0:08:25.282 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Saturday 14 December 2024  18:02:37 -0500 (0:00:00.061)       0:08:25.343 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Saturday 14 December 2024  18:02:38 -0500 (0:00:00.159)       0:08:25.503 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Saturday 14 December 2024  18:02:38 -0500 (0:00:00.059)       0:08:25.562 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}
lsrpackages: cryptsetup

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Saturday 14 December 2024  18:02:38 -0500 (0:00:00.750)       0:08:26.312 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Saturday 14 December 2024  18:02:38 -0500 (0:00:00.061)       0:08:26.374 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.044)       0:08:26.419 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.063)       0:08:26.483 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.052)       0:08:26.535 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.046)       0:08:26.581 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.038)       0:08:26.620 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.037)       0:08:26.658 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.036)       0:08:26.694 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [], 
        "_storage_test_expected_crypttab_entries": "0", 
        "_storage_test_expected_crypttab_key_file": "-"
    }, 
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.046)       0:08:26.741 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.044)       0:08:26.786 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.040)       0:08:26.826 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.038)       0:08:26.865 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.037)       0:08:26.902 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null, 
        "_storage_test_expected_crypttab_entries": null, 
        "_storage_test_expected_crypttab_key_file": null
    }, 
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.037)       0:08:26.940 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.045)       0:08:26.985 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.050)       0:08:27.035 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.062)       0:08:27.098 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.057)       0:08:27.155 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.045)       0:08:27.201 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.069)       0:08:27.271 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.063)       0:08:27.334 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Saturday 14 December 2024  18:02:39 -0500 (0:00:00.045)       0:08:27.380 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Saturday 14 December 2024  18:02:40 -0500 (0:00:00.050)       0:08:27.431 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Saturday 14 December 2024  18:02:40 -0500 (0:00:00.047)       0:08:27.478 ***** 
ok: [managed-node2] => {
    "bytes": 4294967296, 
    "changed": false, 
    "lvm": "4g", 
    "parted": "4GiB", 
    "size": "4 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Saturday 14 December 2024  18:02:40 -0500 (0:00:00.330)       0:08:27.809 ***** 
ok: [managed-node2] => {
    "bytes": 4294967296, 
    "changed": false, 
    "lvm": "4g", 
    "parted": "4GiB", 
    "size": "4 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Saturday 14 December 2024  18:02:40 -0500 (0:00:00.387)       0:08:28.196 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_expected_size": "4294967296"
    }, 
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Saturday 14 December 2024  18:02:40 -0500 (0:00:00.077)       0:08:28.273 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Saturday 14 December 2024  18:02:40 -0500 (0:00:00.083)       0:08:28.357 ***** 
ok: [managed-node2] => {
    "bytes": 10737418240, 
    "changed": false, 
    "lvm": "10g", 
    "parted": "10GiB", 
    "size": "10 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Saturday 14 December 2024  18:02:41 -0500 (0:00:00.371)       0:08:28.728 ***** 
skipping: [managed-node2] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Saturday 14 December 2024  18:02:41 -0500 (0:00:00.048)       0:08:28.776 ***** 
skipping: [managed-node2] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Saturday 14 December 2024  18:02:41 -0500 (0:00:00.053)       0:08:28.830 ***** 
skipping: [managed-node2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Saturday 14 December 2024  18:02:41 -0500 (0:00:00.066)       0:08:28.897 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Saturday 14 December 2024  18:02:41 -0500 (0:00:00.066)       0:08:28.963 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Saturday 14 December 2024  18:02:41 -0500 (0:00:00.059)       0:08:29.022 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Saturday 14 December 2024  18:02:41 -0500 (0:00:00.056)       0:08:29.079 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Saturday 14 December 2024  18:02:41 -0500 (0:00:00.056)       0:08:29.135 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Saturday 14 December 2024  18:02:41 -0500 (0:00:00.060)       0:08:29.196 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Saturday 14 December 2024  18:02:41 -0500 (0:00:00.063)       0:08:29.259 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Saturday 14 December 2024  18:02:41 -0500 (0:00:00.061)       0:08:29.321 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Saturday 14 December 2024  18:02:42 -0500 (0:00:00.063)       0:08:29.384 ***** 
skipping: [managed-node2] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Saturday 14 December 2024  18:02:42 -0500 (0:00:00.048)       0:08:29.433 ***** 
skipping: [managed-node2] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Saturday 14 December 2024  18:02:42 -0500 (0:00:00.062)       0:08:29.495 ***** 
skipping: [managed-node2] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Saturday 14 December 2024  18:02:42 -0500 (0:00:00.061)       0:08:29.556 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Saturday 14 December 2024  18:02:42 -0500 (0:00:00.062)       0:08:29.619 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Saturday 14 December 2024  18:02:42 -0500 (0:00:00.058)       0:08:29.677 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Saturday 14 December 2024  18:02:42 -0500 (0:00:00.043)       0:08:29.721 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Saturday 14 December 2024  18:02:42 -0500 (0:00:00.046)       0:08:29.767 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Saturday 14 December 2024  18:02:42 -0500 (0:00:00.036)       0:08:29.804 ***** 
ok: [managed-node2] => {
    "storage_test_actual_size": {
        "bytes": 4294967296, 
        "changed": false, 
        "failed": false, 
        "lvm": "4g", 
        "parted": "4GiB", 
        "size": "4 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Saturday 14 December 2024  18:02:42 -0500 (0:00:00.044)       0:08:29.848 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Saturday 14 December 2024  18:02:42 -0500 (0:00:00.043)       0:08:29.892 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Saturday 14 December 2024  18:02:42 -0500 (0:00:00.051)       0:08:29.943 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "lvs", 
        "--noheadings", 
        "--nameprefixes", 
        "--units=b", 
        "--nosuffix", 
        "--unquoted", 
        "-o", 
        "name,attr,cache_total_blocks,chunk_size,segtype", 
        "foo/test1"
    ], 
    "delta": "0:00:00.017614", 
    "end": "2024-12-14 18:02:42.841643", 
    "rc": 0, 
    "start": "2024-12-14 18:02:42.824029"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Saturday 14 December 2024  18:02:42 -0500 (0:00:00.346)       0:08:30.289 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    }, 
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Saturday 14 December 2024  18:02:42 -0500 (0:00:00.047)       0:08:30.336 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Saturday 14 December 2024  18:02:43 -0500 (0:00:00.054)       0:08:30.390 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Saturday 14 December 2024  18:02:43 -0500 (0:00:00.048)       0:08:30.439 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Saturday 14 December 2024  18:02:43 -0500 (0:00:00.040)       0:08:30.479 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Saturday 14 December 2024  18:02:43 -0500 (0:00:00.039)       0:08:30.519 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Saturday 14 December 2024  18:02:43 -0500 (0:00:00.042)       0:08:30.562 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    }, 
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Saturday 14 December 2024  18:02:43 -0500 (0:00:00.039)       0:08:30.601 ***** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Saturday 14 December 2024  18:02:43 -0500 (0:00:00.035)       0:08:30.636 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null, 
        "storage_test_crypttab": null, 
        "storage_test_fstab": null
    }, 
    "changed": false
}

TASK [Create a file] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml:12
Saturday 14 December 2024  18:02:43 -0500 (0:00:00.049)       0:08:30.686 ***** 
changed: [managed-node2] => {
    "changed": true, 
    "dest": "/opt/test1/quux", 
    "gid": 0, 
    "group": "root", 
    "mode": "0644", 
    "owner": "root", 
    "secontext": "unconfined_u:object_r:unlabeled_t:s0", 
    "size": 0, 
    "state": "file", 
    "uid": 0
}

TASK [Test for correct handling of safe_mode] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:468
Saturday 14 December 2024  18:02:43 -0500 (0:00:00.515)       0:08:31.202 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node2

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Saturday 14 December 2024  18:02:43 -0500 (0:00:00.172)       0:08:31.375 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_pools_global": [], 
        "storage_safe_mode_global": true, 
        "storage_volumes_global": []
    }, 
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Saturday 14 December 2024  18:02:44 -0500 (0:00:00.094)       0:08:31.469 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  18:02:44 -0500 (0:00:00.196)       0:08:31.666 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  18:02:44 -0500 (0:00:00.079)       0:08:31.745 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  18:02:44 -0500 (0:00:00.082)       0:08:31.827 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  18:02:44 -0500 (0:00:00.154)       0:08:31.982 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  18:02:44 -0500 (0:00:00.057)       0:08:32.039 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  18:02:44 -0500 (0:00:00.128)       0:08:32.168 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  18:02:44 -0500 (0:00:00.107)       0:08:32.276 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  18:02:44 -0500 (0:00:00.089)       0:08:32.365 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  18:02:45 -0500 (0:00:00.151)       0:08:32.517 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  18:02:46 -0500 (0:00:01.285)       0:08:33.802 ***** 
ok: [managed-node2] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ], 
            "name": "foo", 
            "type": "lvm", 
            "volumes": [
                {
                    "encryption": true, 
                    "encryption_password": "yabbadabbadoo", 
                    "mount_point": "/opt/test1", 
                    "name": "test1", 
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  18:02:46 -0500 (0:00:00.073)       0:08:33.876 ***** 
ok: [managed-node2] => {
    "storage_volumes": []
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  18:02:46 -0500 (0:00:00.068)       0:08:33.944 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [
        "cryptsetup", 
        "lvm2"
    ], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  18:02:50 -0500 (0:00:04.196)       0:08:38.141 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  18:02:50 -0500 (0:00:00.163)       0:08:38.304 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  18:02:50 -0500 (0:00:00.060)       0:08:38.365 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  18:02:51 -0500 (0:00:00.082)       0:08:38.447 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  18:02:51 -0500 (0:00:00.106)       0:08:38.554 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed", 
        "7:lvm2-2.02.187-6.el7_9.5.x86_64 providing lvm2 is already installed", 
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: cryptsetup kpartx lvm2

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  18:02:52 -0500 (0:00:01.042)       0:08:39.596 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service": {
                "name": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  18:02:53 -0500 (0:00:01.181)       0:08:40.778 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": [
            "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service"
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  18:02:53 -0500 (0:00:00.087)       0:08:40.865 ***** 
changed: [managed-node2] => (item=systemd-cryptsetup@luks\x2d3365bc93\x2d2fc0\x2d47e8\x2db784\x2db63a2126f78c.service) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
    "name": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "After": "systemd-readahead-collect.service systemd-readahead-replay.service dev-mapper-foo\\x2dtest1.device system-systemd\\x2dcryptsetup.slice cryptsetup-pre.target systemd-journald.socket", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "Before": "umount.target cryptsetup.target", 
        "BindsTo": "dev-mapper-foo\\x2dtest1.device", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "Conflicts": "umount.target", 
        "ControlPID": "0", 
        "DefaultDependencies": "no", 
        "Delegate": "no", 
        "Description": "Cryptography Setup for luks-3365bc93-2fc0-47e8-b784-b63a2126f78c", 
        "DevicePolicy": "auto", 
        "Documentation": "man:crypttab(5) man:systemd-cryptsetup-generator(8) man:systemd-cryptsetup@.service(8)", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-3365bc93-2fc0-47e8-b784-b63a2126f78c /dev/mapper/foo-test1 -  ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-3365bc93-2fc0-47e8-b784-b63a2126f78c ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
        "IgnoreOnIsolate": "yes", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "4096", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "loaded", 
        "MainPID": "0", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
        "NeedDaemonReload": "yes", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "yes", 
        "RequiredBy": "cryptsetup.target", 
        "Requires": "system-systemd\\x2dcryptsetup.slice", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "Slice": "system-systemd\\x2dcryptsetup.slice", 
        "SourcePath": "/etc/crypttab", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "journal", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "0", 
        "TimeoutStopUSec": "0", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "Type": "oneshot", 
        "UMask": "0022", 
        "UnitFilePreset": "disabled", 
        "UnitFileState": "bad", 
        "WantedBy": "dev-mapper-foo\\x2dtest1.device", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  18:02:54 -0500 (0:00:00.593)       0:08:41.458 ***** 
fatal: [managed-node2]: FAILED! => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [], 
    "pools": [], 
    "volumes": []
}

MSG:

cannot remove existing formatting on device 'test1' in safe mode due to adding encryption

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Saturday 14 December 2024  18:02:58 -0500 (0:00:04.331)       0:08:45.790 ***** 
fatal: [managed-node2]: FAILED! => {
    "changed": false
}

MSG:

{u'_ansible_no_log': False, u'crypts': [], u'pools': [], u'leaves': [], u'changed': False, u'actions': [], u'failed': True, u'volumes': [], u'invocation': {u'module_args': {u'packages_only': False, u'disklabel_type': None, u'diskvolume_mkfs_option_map': {u'ext4': u'-F', u'ext3': u'-F', u'ext2': u'-F'}, u'safe_mode': True, u'pools': [{u'raid_metadata_version': None, u'encryption_key_size': None, u'encryption_key': None, u'encryption_luks_version': None, u'encryption_tang_url': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_tang_thumbprint': None, u'name': u'foo', u'encryption_password': None, u'encryption': False, u'disks': [u'sda'], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [{u'raid_metadata_version': None, u'mount_device_identifier': u'uuid', u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': u'4g', u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'encryption': True, u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'type': u'lvm', u'encryption_key_size': None, u'deduplication': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'mount_passno': 0, u'mount_user': None, u'raid_spare_count': None, u'name': u'test1', u'cache_mode': None, u'raid_disks': [], u'mount_group': None, u'fs_overwrite_existing': True, u'disks': [u'sda'], u'cached': False, u'thin_pool_size': None, u'thin': False, u'mount_check': 0, u'cache_size': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}], u'shared': False, u'encryption_clevis_pin': None, u'type': u'lvm', u'encryption_cipher': None, u'raid_chunk_size': None}], u'volumes': [], u'pool_defaults': {u'raid_metadata_version': None, u'encryption_cipher': None, u'encryption_key': None, u'encryption_luks_version': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_password': None, u'encryption': False, u'disks': [], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [], u'shared': False, u'type': u'lvm', u'encryption_key_size': None, u'raid_chunk_size': None}, u'volume_defaults': {u'raid_metadata_version': None, u'raid_level': None, u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': 0, u'mount_point': u'', u'compression': None, u'encryption_password': None, u'encryption': False, u'mount_device_identifier': u'uuid', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'thin_pool_name': None, u'type': u'lvm', u'encryption_key_size': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'cache_size': 0, u'raid_spare_count': None, u'cache_mode': None, u'deduplication': None, u'cached': False, u'fs_overwrite_existing': True, u'disks': [], u'thin_pool_size': None, u'thin': None, u'mount_check': 0, u'mount_passno': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}, u'use_partitions': None}}, u'mounts': [], u'packages': [], u'msg': u"cannot remove existing formatting on device 'test1' in safe mode due to adding encryption"}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  18:02:58 -0500 (0:00:00.084)       0:08:45.874 ***** 
changed: [managed-node2] => (item=systemd-cryptsetup@luks\x2d3365bc93\x2d2fc0\x2d47e8\x2db784\x2db63a2126f78c.service) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
    "name": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "ControlPID": "0", 
        "DefaultDependencies": "yes", 
        "Delegate": "no", 
        "Description": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
        "DevicePolicy": "auto", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/dev/null", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
        "IgnoreOnIsolate": "no", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "1048576", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "masked", 
        "MainPID": "0", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "systemd-cryptsetup@luks\\x2d3365bc93\\x2d2fc0\\x2d47e8\\x2db784\\x2db63a2126f78c.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "no", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "inherit", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "1min 30s", 
        "TimeoutStopUSec": "1min 30s", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "UMask": "0022", 
        "UnitFileState": "bad", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Saturday 14 December 2024  18:02:59 -0500 (0:00:00.538)       0:08:46.413 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Saturday 14 December 2024  18:02:59 -0500 (0:00:00.062)       0:08:46.476 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Saturday 14 December 2024  18:02:59 -0500 (0:00:00.080)       0:08:46.556 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Stat the file] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:11
Saturday 14 December 2024  18:02:59 -0500 (0:00:00.058)       0:08:46.615 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217363.709653, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", 
        "ctime": 1734217363.709653, 
        "dev": 64768, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 67, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/x-empty", 
        "mode": "0644", 
        "mtime": 1734217363.709653, 
        "nlink": 1, 
        "path": "/opt/test1/quux", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": "623518059", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Assert file presence] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:16
Saturday 14 December 2024  18:02:59 -0500 (0:00:00.355)       0:08:46.970 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Add encryption to the volume] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:491
Saturday 14 December 2024  18:02:59 -0500 (0:00:00.050)       0:08:47.021 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  18:02:59 -0500 (0:00:00.130)       0:08:47.152 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  18:02:59 -0500 (0:00:00.074)       0:08:47.226 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  18:02:59 -0500 (0:00:00.070)       0:08:47.297 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  18:03:00 -0500 (0:00:00.124)       0:08:47.421 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  18:03:00 -0500 (0:00:00.134)       0:08:47.555 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  18:03:00 -0500 (0:00:00.041)       0:08:47.597 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  18:03:00 -0500 (0:00:00.037)       0:08:47.635 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  18:03:00 -0500 (0:00:00.042)       0:08:47.677 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  18:03:00 -0500 (0:00:00.088)       0:08:47.765 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  18:03:01 -0500 (0:00:01.244)       0:08:49.010 ***** 
ok: [managed-node2] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ], 
            "name": "foo", 
            "type": "lvm", 
            "volumes": [
                {
                    "encryption": true, 
                    "encryption_password": "yabbadabbadoo", 
                    "mount_point": "/opt/test1", 
                    "name": "test1", 
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  18:03:01 -0500 (0:00:00.114)       0:08:49.125 ***** 
ok: [managed-node2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  18:03:01 -0500 (0:00:00.123)       0:08:49.249 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [
        "cryptsetup", 
        "lvm2"
    ], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  18:03:06 -0500 (0:00:04.398)       0:08:53.647 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  18:03:06 -0500 (0:00:00.185)       0:08:53.833 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  18:03:06 -0500 (0:00:00.097)       0:08:53.930 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  18:03:06 -0500 (0:00:00.071)       0:08:54.001 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  18:03:06 -0500 (0:00:00.098)       0:08:54.100 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed", 
        "7:lvm2-2.02.187-6.el7_9.5.x86_64 providing lvm2 is already installed", 
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: cryptsetup kpartx lvm2

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  18:03:07 -0500 (0:00:01.181)       0:08:55.281 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  18:03:09 -0500 (0:00:01.299)       0:08:56.581 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  18:03:09 -0500 (0:00:00.161)       0:08:56.742 ***** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  18:03:09 -0500 (0:00:00.064)       0:08:56.807 ***** 
changed: [managed-node2] => {
    "actions": [
        {
            "action": "destroy format", 
            "device": "/dev/mapper/foo-test1", 
            "fs_type": "xfs"
        }, 
        {
            "action": "create format", 
            "device": "/dev/mapper/foo-test1", 
            "fs_type": "luks"
        }, 
        {
            "action": "create device", 
            "device": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
            "fs_type": null
        }, 
        {
            "action": "create format", 
            "device": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
            "fs_type": "xfs"
        }
    ], 
    "changed": true, 
    "crypts": [
        {
            "backing_device": "/dev/mapper/foo-test1", 
            "name": "luks-4ce986c5-0285-435c-8ffc-f915139facba", 
            "password": "-", 
            "state": "present"
        }
    ], 
    "leaves": [
        "/dev/sdb", 
        "/dev/sdc", 
        "/dev/sdd", 
        "/dev/sde", 
        "/dev/sdf", 
        "/dev/sdg", 
        "/dev/sdh", 
        "/dev/sdi", 
        "/dev/xvda1", 
        "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba"
    ], 
    "mounts": [
        {
            "fstype": "xfs", 
            "path": "/opt/test1", 
            "src": "/dev/mapper/foo-test1", 
            "state": "absent"
        }, 
        {
            "dump": 0, 
            "fstype": "xfs", 
            "group": null, 
            "mode": null, 
            "opts": "defaults", 
            "owner": null, 
            "passno": 0, 
            "path": "/opt/test1", 
            "src": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
            "state": "mounted"
        }
    ], 
    "packages": [
        "cryptsetup", 
        "xfsprogs", 
        "e2fsprogs", 
        "lvm2"
    ], 
    "pools": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_cipher": null, 
            "encryption_clevis_pin": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": null, 
            "encryption_tang_thumbprint": null, 
            "encryption_tang_url": null, 
            "grow_to_fill": false, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "shared": false, 
            "state": "present", 
            "type": "lvm", 
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
                    "_kernel_device": "/dev/dm-1", 
                    "_mount_id": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
                    "_raw_device": "/dev/mapper/foo-test1", 
                    "_raw_kernel_device": "/dev/dm-0", 
                    "cache_devices": [], 
                    "cache_mode": null, 
                    "cache_size": 0, 
                    "cached": false, 
                    "compression": null, 
                    "deduplication": null, 
                    "disks": [
                        "sda"
                    ], 
                    "encryption": true, 
                    "encryption_cipher": null, 
                    "encryption_key": null, 
                    "encryption_key_size": null, 
                    "encryption_luks_version": null, 
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                    "fs_create_options": "", 
                    "fs_label": "", 
                    "fs_overwrite_existing": true, 
                    "fs_type": "xfs", 
                    "mount_check": 0, 
                    "mount_device_identifier": "uuid", 
                    "mount_group": null, 
                    "mount_mode": null, 
                    "mount_options": "defaults", 
                    "mount_passno": 0, 
                    "mount_point": "/opt/test1", 
                    "mount_user": null, 
                    "name": "test1", 
                    "raid_chunk_size": null, 
                    "raid_device_count": null, 
                    "raid_disks": [], 
                    "raid_level": null, 
                    "raid_metadata_version": null, 
                    "raid_spare_count": null, 
                    "raid_stripe_size": null, 
                    "size": "4g", 
                    "state": "present", 
                    "thin": false, 
                    "thin_pool_name": null, 
                    "thin_pool_size": null, 
                    "type": "lvm", 
                    "vdo_pool_size": null
                }
            ]
        }
    ], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Saturday 14 December 2024  18:03:20 -0500 (0:00:10.881)       0:09:07.688 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Saturday 14 December 2024  18:03:20 -0500 (0:00:00.058)       0:09:07.747 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217344.9426613, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "3fceedeef6c619b69ada96279531b69ed89734ba", 
        "ctime": 1734217344.9386613, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 263588, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0644", 
        "mtime": 1734217344.9386613, 
        "nlink": 1, 
        "path": "/etc/fstab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 1279, 
        "uid": 0, 
        "version": "606402765", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Saturday 14 December 2024  18:03:20 -0500 (0:00:00.386)       0:09:08.133 ***** 
ok: [managed-node2] => {
    "backup": "", 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  18:03:21 -0500 (0:00:00.438)       0:09:08.572 ***** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Saturday 14 December 2024  18:03:21 -0500 (0:00:00.068)       0:09:08.640 ***** 
ok: [managed-node2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format", 
                "device": "/dev/mapper/foo-test1", 
                "fs_type": "xfs"
            }, 
            {
                "action": "create format", 
                "device": "/dev/mapper/foo-test1", 
                "fs_type": "luks"
            }, 
            {
                "action": "create device", 
                "device": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
                "fs_type": null
            }, 
            {
                "action": "create format", 
                "device": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
                "fs_type": "xfs"
            }
        ], 
        "changed": true, 
        "crypts": [
            {
                "backing_device": "/dev/mapper/foo-test1", 
                "name": "luks-4ce986c5-0285-435c-8ffc-f915139facba", 
                "password": "-", 
                "state": "present"
            }
        ], 
        "failed": false, 
        "leaves": [
            "/dev/sdb", 
            "/dev/sdc", 
            "/dev/sdd", 
            "/dev/sde", 
            "/dev/sdf", 
            "/dev/sdg", 
            "/dev/sdh", 
            "/dev/sdi", 
            "/dev/xvda1", 
            "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba"
        ], 
        "mounts": [
            {
                "fstype": "xfs", 
                "path": "/opt/test1", 
                "src": "/dev/mapper/foo-test1", 
                "state": "absent"
            }, 
            {
                "dump": 0, 
                "fstype": "xfs", 
                "group": null, 
                "mode": null, 
                "opts": "defaults", 
                "owner": null, 
                "passno": 0, 
                "path": "/opt/test1", 
                "src": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
                "state": "mounted"
            }
        ], 
        "packages": [
            "cryptsetup", 
            "xfsprogs", 
            "e2fsprogs", 
            "lvm2"
        ], 
        "pools": [
            {
                "disks": [
                    "sda"
                ], 
                "encryption": false, 
                "encryption_cipher": null, 
                "encryption_clevis_pin": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": null, 
                "encryption_tang_thumbprint": null, 
                "encryption_tang_url": null, 
                "grow_to_fill": false, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "shared": false, 
                "state": "present", 
                "type": "lvm", 
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
                        "_kernel_device": "/dev/dm-1", 
                        "_mount_id": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
                        "_raw_device": "/dev/mapper/foo-test1", 
                        "_raw_kernel_device": "/dev/dm-0", 
                        "cache_devices": [], 
                        "cache_mode": null, 
                        "cache_size": 0, 
                        "cached": false, 
                        "compression": null, 
                        "deduplication": null, 
                        "disks": [
                            "sda"
                        ], 
                        "encryption": true, 
                        "encryption_cipher": null, 
                        "encryption_key": null, 
                        "encryption_key_size": null, 
                        "encryption_luks_version": null, 
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                        "fs_create_options": "", 
                        "fs_label": "", 
                        "fs_overwrite_existing": true, 
                        "fs_type": "xfs", 
                        "mount_check": 0, 
                        "mount_device_identifier": "uuid", 
                        "mount_group": null, 
                        "mount_mode": null, 
                        "mount_options": "defaults", 
                        "mount_passno": 0, 
                        "mount_point": "/opt/test1", 
                        "mount_user": null, 
                        "name": "test1", 
                        "raid_chunk_size": null, 
                        "raid_device_count": null, 
                        "raid_disks": [], 
                        "raid_level": null, 
                        "raid_metadata_version": null, 
                        "raid_spare_count": null, 
                        "raid_stripe_size": null, 
                        "size": "4g", 
                        "state": "present", 
                        "thin": false, 
                        "thin_pool_name": null, 
                        "thin_pool_size": null, 
                        "type": "lvm", 
                        "vdo_pool_size": null
                    }
                ]
            }
        ], 
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Saturday 14 December 2024  18:03:21 -0500 (0:00:00.111)       0:09:08.752 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ], 
                "encryption": false, 
                "encryption_cipher": null, 
                "encryption_clevis_pin": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": null, 
                "encryption_tang_thumbprint": null, 
                "encryption_tang_url": null, 
                "grow_to_fill": false, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "shared": false, 
                "state": "present", 
                "type": "lvm", 
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
                        "_kernel_device": "/dev/dm-1", 
                        "_mount_id": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
                        "_raw_device": "/dev/mapper/foo-test1", 
                        "_raw_kernel_device": "/dev/dm-0", 
                        "cache_devices": [], 
                        "cache_mode": null, 
                        "cache_size": 0, 
                        "cached": false, 
                        "compression": null, 
                        "deduplication": null, 
                        "disks": [
                            "sda"
                        ], 
                        "encryption": true, 
                        "encryption_cipher": null, 
                        "encryption_key": null, 
                        "encryption_key_size": null, 
                        "encryption_luks_version": null, 
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                        "fs_create_options": "", 
                        "fs_label": "", 
                        "fs_overwrite_existing": true, 
                        "fs_type": "xfs", 
                        "mount_check": 0, 
                        "mount_device_identifier": "uuid", 
                        "mount_group": null, 
                        "mount_mode": null, 
                        "mount_options": "defaults", 
                        "mount_passno": 0, 
                        "mount_point": "/opt/test1", 
                        "mount_user": null, 
                        "name": "test1", 
                        "raid_chunk_size": null, 
                        "raid_device_count": null, 
                        "raid_disks": [], 
                        "raid_level": null, 
                        "raid_metadata_version": null, 
                        "raid_spare_count": null, 
                        "raid_stripe_size": null, 
                        "size": "4g", 
                        "state": "present", 
                        "thin": false, 
                        "thin_pool_name": null, 
                        "thin_pool_size": null, 
                        "type": "lvm", 
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Saturday 14 December 2024  18:03:21 -0500 (0:00:00.104)       0:09:08.857 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Saturday 14 December 2024  18:03:21 -0500 (0:00:00.067)       0:09:08.924 ***** 
changed: [managed-node2] => (item={u'src': u'/dev/mapper/foo-test1', u'state': u'absent', u'fstype': u'xfs', u'path': u'/opt/test1'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": true, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "fstype": "xfs", 
        "path": "/opt/test1", 
        "src": "/dev/mapper/foo-test1", 
        "state": "absent"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Saturday 14 December 2024  18:03:22 -0500 (0:00:00.489)       0:09:09.414 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Saturday 14 December 2024  18:03:22 -0500 (0:00:00.677)       0:09:10.092 ***** 
changed: [managed-node2] => (item={u'src': u'/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": true, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
        "state": "mounted"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Saturday 14 December 2024  18:03:23 -0500 (0:00:00.434)       0:09:10.527 ***** 
skipping: [managed-node2] => (item={u'src': u'/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'xfs', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'})  => {
    "ansible_loop_var": "mount_info", 
    "changed": false, 
    "mount_info": {
        "dump": 0, 
        "fstype": "xfs", 
        "group": null, 
        "mode": null, 
        "opts": "defaults", 
        "owner": null, 
        "passno": 0, 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
        "state": "mounted"
    }, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Saturday 14 December 2024  18:03:23 -0500 (0:00:00.090)       0:09:10.617 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Saturday 14 December 2024  18:03:23 -0500 (0:00:00.624)       0:09:11.242 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217348.6406596, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", 
        "ctime": 1734217346.4286606, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 917510, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/x-empty", 
        "mode": "0600", 
        "mtime": 1734217346.4286606, 
        "nlink": 1, 
        "path": "/etc/crypttab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": false, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": "606409930", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Saturday 14 December 2024  18:03:24 -0500 (0:00:00.538)       0:09:11.780 ***** 
changed: [managed-node2] => (item={u'state': u'present', u'password': u'-', u'name': u'luks-4ce986c5-0285-435c-8ffc-f915139facba', u'backing_device': u'/dev/mapper/foo-test1'}) => {
    "ansible_loop_var": "entry", 
    "backup": "", 
    "changed": true, 
    "entry": {
        "backing_device": "/dev/mapper/foo-test1", 
        "name": "luks-4ce986c5-0285-435c-8ffc-f915139facba", 
        "password": "-", 
        "state": "present"
    }
}

MSG:

line added

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Saturday 14 December 2024  18:03:24 -0500 (0:00:00.497)       0:09:12.278 ***** 
ok: [managed-node2]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:507
Saturday 14 December 2024  18:03:25 -0500 (0:00:00.802)       0:09:13.081 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node2

TASK [Print out pool information] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Saturday 14 December 2024  18:03:25 -0500 (0:00:00.149)       0:09:13.230 ***** 
ok: [managed-node2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_cipher": null, 
            "encryption_clevis_pin": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": null, 
            "encryption_tang_thumbprint": null, 
            "encryption_tang_url": null, 
            "grow_to_fill": false, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "shared": false, 
            "state": "present", 
            "type": "lvm", 
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
                    "_kernel_device": "/dev/dm-1", 
                    "_mount_id": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
                    "_raw_device": "/dev/mapper/foo-test1", 
                    "_raw_kernel_device": "/dev/dm-0", 
                    "cache_devices": [], 
                    "cache_mode": null, 
                    "cache_size": 0, 
                    "cached": false, 
                    "compression": null, 
                    "deduplication": null, 
                    "disks": [
                        "sda"
                    ], 
                    "encryption": true, 
                    "encryption_cipher": null, 
                    "encryption_key": null, 
                    "encryption_key_size": null, 
                    "encryption_luks_version": null, 
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", 
                    "fs_create_options": "", 
                    "fs_label": "", 
                    "fs_overwrite_existing": true, 
                    "fs_type": "xfs", 
                    "mount_check": 0, 
                    "mount_device_identifier": "uuid", 
                    "mount_group": null, 
                    "mount_mode": null, 
                    "mount_options": "defaults", 
                    "mount_passno": 0, 
                    "mount_point": "/opt/test1", 
                    "mount_user": null, 
                    "name": "test1", 
                    "raid_chunk_size": null, 
                    "raid_device_count": null, 
                    "raid_disks": [], 
                    "raid_level": null, 
                    "raid_metadata_version": null, 
                    "raid_spare_count": null, 
                    "raid_stripe_size": null, 
                    "size": "4g", 
                    "state": "present", 
                    "thin": false, 
                    "thin_pool_name": null, 
                    "thin_pool_size": null, 
                    "type": "lvm", 
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Saturday 14 December 2024  18:03:25 -0500 (0:00:00.138)       0:09:13.369 ***** 
skipping: [managed-node2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Saturday 14 December 2024  18:03:26 -0500 (0:00:00.093)       0:09:13.462 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "crypto_LUKS", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/mapper/foo-test1", 
            "size": "4G", 
            "type": "lvm", 
            "uuid": "4ce986c5-0285-435c-8ffc-f915139facba"
        }, 
        "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba": {
            "fstype": "xfs", 
            "label": "", 
            "mountpoint": "/opt/test1", 
            "name": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
            "size": "4G", 
            "type": "crypt", 
            "uuid": "4d6130e5-aad0-4567-9b1b-53f4614778b0"
        }, 
        "/dev/sda": {
            "fstype": "LVM2_member", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sda", 
            "size": "10G", 
            "type": "disk", 
            "uuid": "xa7q16-X4RQ-vFWY-Ftcj-AFkA-0I7B-z0LDg5"
        }, 
        "/dev/sdb": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdb", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdc": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdc", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdd": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdd", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sde": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sde", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdf": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdf", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdg": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdg", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdh": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdh", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdi": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdi", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/xvda", 
            "size": "250G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda1": {
            "fstype": "ext4", 
            "label": "", 
            "mountpoint": "/", 
            "name": "/dev/xvda1", 
            "size": "250G", 
            "type": "partition", 
            "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Saturday 14 December 2024  18:03:26 -0500 (0:00:00.626)       0:09:14.088 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/fstab"
    ], 
    "delta": "0:00:00.002947", 
    "end": "2024-12-14 18:03:27.026485", 
    "rc": 0, 
    "start": "2024-12-14 18:03:27.023538"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Thu Jun 20 10:23:46 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 /                       ext4    defaults        1 1
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Saturday 14 December 2024  18:03:27 -0500 (0:00:00.424)       0:09:14.513 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/crypttab"
    ], 
    "delta": "0:00:00.002837", 
    "end": "2024-12-14 18:03:27.534894", 
    "failed_when_result": false, 
    "rc": 0, 
    "start": "2024-12-14 18:03:27.532057"
}

STDOUT:

luks-4ce986c5-0285-435c-8ffc-f915139facba /dev/mapper/foo-test1 -

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Saturday 14 December 2024  18:03:27 -0500 (0:00:00.498)       0:09:15.012 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node2

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Saturday 14 December 2024  18:03:27 -0500 (0:00:00.148)       0:09:15.161 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members", 
            "volumes"
        ]
    }, 
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Saturday 14 December 2024  18:03:27 -0500 (0:00:00.066)       0:09:15.228 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "vgs", 
        "--noheadings", 
        "--binary", 
        "-o", 
        "shared", 
        "foo"
    ], 
    "delta": "0:00:00.018051", 
    "end": "2024-12-14 18:03:28.184574", 
    "rc": 0, 
    "start": "2024-12-14 18:03:28.166523"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Saturday 14 December 2024  18:03:28 -0500 (0:00:00.430)       0:09:15.658 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Saturday 14 December 2024  18:03:28 -0500 (0:00:00.079)       0:09:15.738 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node2

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Saturday 14 December 2024  18:03:28 -0500 (0:00:00.156)       0:09:15.894 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1", 
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    }, 
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Saturday 14 December 2024  18:03:28 -0500 (0:00:00.074)       0:09:15.969 ***** 
ok: [managed-node2] => (item=/dev/sda) => {
    "ansible_loop_var": "pv", 
    "changed": false, 
    "device": "/dev/sda", 
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Saturday 14 December 2024  18:03:28 -0500 (0:00:00.370)       0:09:16.339 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    }, 
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Saturday 14 December 2024  18:03:29 -0500 (0:00:00.066)       0:09:16.405 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    }, 
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Saturday 14 December 2024  18:03:29 -0500 (0:00:00.071)       0:09:16.476 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Saturday 14 December 2024  18:03:29 -0500 (0:00:00.072)       0:09:16.549 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    }, 
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Saturday 14 December 2024  18:03:29 -0500 (0:00:00.066)       0:09:16.616 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    }, 
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Saturday 14 December 2024  18:03:29 -0500 (0:00:00.067)       0:09:16.683 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Saturday 14 December 2024  18:03:29 -0500 (0:00:00.056)       0:09:16.740 ***** 
ok: [managed-node2] => (item=/dev/sda) => {
    "ansible_loop_var": "pv", 
    "changed": false, 
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Saturday 14 December 2024  18:03:29 -0500 (0:00:00.081)       0:09:16.821 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.8.171 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Saturday 14 December 2024  18:03:29 -0500 (0:00:00.305)       0:09:17.127 ***** 
skipping: [managed-node2] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv", 
    "changed": false, 
    "skip_reason": "Conditional result was False", 
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Saturday 14 December 2024  18:03:29 -0500 (0:00:00.051)       0:09:17.179 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node2

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Saturday 14 December 2024  18:03:29 -0500 (0:00:00.079)       0:09:17.258 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Saturday 14 December 2024  18:03:29 -0500 (0:00:00.039)       0:09:17.297 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Saturday 14 December 2024  18:03:29 -0500 (0:00:00.039)       0:09:17.337 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Saturday 14 December 2024  18:03:29 -0500 (0:00:00.038)       0:09:17.375 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Saturday 14 December 2024  18:03:30 -0500 (0:00:00.038)       0:09:17.414 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Saturday 14 December 2024  18:03:30 -0500 (0:00:00.043)       0:09:17.457 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Saturday 14 December 2024  18:03:30 -0500 (0:00:00.049)       0:09:17.507 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Saturday 14 December 2024  18:03:30 -0500 (0:00:00.059)       0:09:17.566 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Saturday 14 December 2024  18:03:30 -0500 (0:00:00.066)       0:09:17.633 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Saturday 14 December 2024  18:03:30 -0500 (0:00:00.069)       0:09:17.702 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Saturday 14 December 2024  18:03:30 -0500 (0:00:00.063)       0:09:17.765 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null, 
        "storage_test_md_chunk_size_re": null, 
        "storage_test_md_metadata_version_re": null, 
        "storage_test_md_spare_devices_re": null
    }, 
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Saturday 14 December 2024  18:03:30 -0500 (0:00:00.061)       0:09:17.827 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Saturday 14 December 2024  18:03:30 -0500 (0:00:00.123)       0:09:17.951 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node2

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Saturday 14 December 2024  18:03:30 -0500 (0:00:00.139)       0:09:18.091 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Saturday 14 December 2024  18:03:30 -0500 (0:00:00.060)       0:09:18.151 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Saturday 14 December 2024  18:03:30 -0500 (0:00:00.059)       0:09:18.211 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Saturday 14 December 2024  18:03:30 -0500 (0:00:00.058)       0:09:18.270 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Saturday 14 December 2024  18:03:30 -0500 (0:00:00.059)       0:09:18.330 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Saturday 14 December 2024  18:03:31 -0500 (0:00:00.065)       0:09:18.395 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Saturday 14 December 2024  18:03:31 -0500 (0:00:00.062)       0:09:18.458 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Saturday 14 December 2024  18:03:31 -0500 (0:00:00.059)       0:09:18.517 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Saturday 14 December 2024  18:03:31 -0500 (0:00:00.123)       0:09:18.640 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node2

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Saturday 14 December 2024  18:03:31 -0500 (0:00:00.205)       0:09:18.845 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Saturday 14 December 2024  18:03:31 -0500 (0:00:00.056)       0:09:18.902 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Saturday 14 December 2024  18:03:31 -0500 (0:00:00.057)       0:09:18.960 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Saturday 14 December 2024  18:03:31 -0500 (0:00:00.057)       0:09:19.017 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    }, 
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Saturday 14 December 2024  18:03:31 -0500 (0:00:00.071)       0:09:19.088 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node2

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Saturday 14 December 2024  18:03:31 -0500 (0:00:00.135)       0:09:19.223 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0", 
        "_storage_test_expected_crypttab_key_file": "-"
    }, 
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Saturday 14 December 2024  18:03:31 -0500 (0:00:00.066)       0:09:19.290 ***** 
skipping: [managed-node2] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda", 
    "ansible_loop_var": "_storage_test_pool_member_path", 
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Saturday 14 December 2024  18:03:31 -0500 (0:00:00.069)       0:09:19.359 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node2

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Saturday 14 December 2024  18:03:32 -0500 (0:00:00.130)       0:09:19.490 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    }, 
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Saturday 14 December 2024  18:03:32 -0500 (0:00:00.111)       0:09:19.601 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Saturday 14 December 2024  18:03:32 -0500 (0:00:00.116)       0:09:19.718 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Saturday 14 December 2024  18:03:32 -0500 (0:00:00.067)       0:09:19.786 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Saturday 14 December 2024  18:03:32 -0500 (0:00:00.062)       0:09:19.848 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Saturday 14 December 2024  18:03:32 -0500 (0:00:00.056)       0:09:19.905 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    }, 
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Saturday 14 December 2024  18:03:32 -0500 (0:00:00.061)       0:09:19.967 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null, 
        "_storage_test_crypttab_key_file": null
    }, 
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Saturday 14 December 2024  18:03:32 -0500 (0:00:00.068)       0:09:20.035 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Saturday 14 December 2024  18:03:32 -0500 (0:00:00.171)       0:09:20.207 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node2

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Saturday 14 December 2024  18:03:32 -0500 (0:00:00.140)       0:09:20.347 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Saturday 14 December 2024  18:03:33 -0500 (0:00:00.066)       0:09:20.414 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Saturday 14 December 2024  18:03:33 -0500 (0:00:00.095)       0:09:20.509 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Saturday 14 December 2024  18:03:33 -0500 (0:00:00.109)       0:09:20.619 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Saturday 14 December 2024  18:03:33 -0500 (0:00:00.080)       0:09:20.700 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Saturday 14 December 2024  18:03:33 -0500 (0:00:00.114)       0:09:20.814 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Saturday 14 December 2024  18:03:33 -0500 (0:00:00.126)       0:09:20.941 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    }, 
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Saturday 14 December 2024  18:03:33 -0500 (0:00:00.075)       0:09:21.016 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node2

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Saturday 14 December 2024  18:03:33 -0500 (0:00:00.167)       0:09:21.184 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Saturday 14 December 2024  18:03:33 -0500 (0:00:00.066)       0:09:21.251 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Saturday 14 December 2024  18:03:33 -0500 (0:00:00.080)       0:09:21.332 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Saturday 14 December 2024  18:03:34 -0500 (0:00:00.060)       0:09:21.393 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Saturday 14 December 2024  18:03:34 -0500 (0:00:00.059)       0:09:21.452 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Saturday 14 December 2024  18:03:34 -0500 (0:00:00.046)       0:09:21.499 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    }, 
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Saturday 14 December 2024  18:03:34 -0500 (0:00:00.040)       0:09:21.539 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null, 
        "_storage_test_expected_pv_count": null, 
        "_storage_test_expected_pv_type": null, 
        "_storage_test_pool_pvs": [], 
        "_storage_test_pool_pvs_lvm": []
    }, 
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Saturday 14 December 2024  18:03:34 -0500 (0:00:00.040)       0:09:21.580 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node2

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Saturday 14 December 2024  18:03:34 -0500 (0:00:00.074)       0:09:21.655 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true, 
        "_storage_volume_tests": [
            "mount", 
            "fstab", 
            "fs", 
            "device", 
            "encryption", 
            "md", 
            "size", 
            "cache"
        ]
    }, 
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Saturday 14 December 2024  18:03:34 -0500 (0:00:00.058)       0:09:21.713 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node2

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Saturday 14 December 2024  18:03:34 -0500 (0:00:00.229)       0:09:21.943 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba"
    }, 
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Saturday 14 December 2024  18:03:34 -0500 (0:00:00.042)       0:09:21.985 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1", 
        "storage_test_swap_expected_matches": "0"
    }, 
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Saturday 14 December 2024  18:03:34 -0500 (0:00:00.045)       0:09:22.031 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Saturday 14 December 2024  18:03:34 -0500 (0:00:00.045)       0:09:22.076 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Saturday 14 December 2024  18:03:34 -0500 (0:00:00.059)       0:09:22.136 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Saturday 14 December 2024  18:03:34 -0500 (0:00:00.042)       0:09:22.179 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Saturday 14 December 2024  18:03:34 -0500 (0:00:00.050)       0:09:22.229 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Saturday 14 December 2024  18:03:34 -0500 (0:00:00.042)       0:09:22.272 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Saturday 14 December 2024  18:03:34 -0500 (0:00:00.051)       0:09:22.323 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Saturday 14 December 2024  18:03:35 -0500 (0:00:00.060)       0:09:22.384 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Saturday 14 December 2024  18:03:35 -0500 (0:00:00.069)       0:09:22.454 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null, 
        "storage_test_mount_expected_mount_point": null, 
        "storage_test_swap_expected_matches": null, 
        "storage_test_swaps": null, 
        "storage_test_sys_node": null
    }, 
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Saturday 14 December 2024  18:03:35 -0500 (0:00:00.067)       0:09:22.522 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1", 
        "storage_test_fstab_expected_mount_options_matches": "1", 
        "storage_test_fstab_expected_mount_point_matches": "1", 
        "storage_test_fstab_id_matches": [
            "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba "
        ], 
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ], 
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    }, 
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Saturday 14 December 2024  18:03:35 -0500 (0:00:00.106)       0:09:22.629 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Saturday 14 December 2024  18:03:35 -0500 (0:00:00.251)       0:09:22.880 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Saturday 14 December 2024  18:03:35 -0500 (0:00:00.072)       0:09:22.953 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Saturday 14 December 2024  18:03:35 -0500 (0:00:00.082)       0:09:23.036 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Saturday 14 December 2024  18:03:35 -0500 (0:00:00.074)       0:09:23.110 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null, 
        "storage_test_fstab_expected_mount_options_matches": null, 
        "storage_test_fstab_expected_mount_point_matches": null, 
        "storage_test_fstab_id_matches": null, 
        "storage_test_fstab_mount_options_matches": null, 
        "storage_test_fstab_mount_point_matches": null
    }, 
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Saturday 14 December 2024  18:03:35 -0500 (0:00:00.063)       0:09:23.173 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Saturday 14 December 2024  18:03:35 -0500 (0:00:00.090)       0:09:23.264 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Saturday 14 December 2024  18:03:35 -0500 (0:00:00.094)       0:09:23.359 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217400.0496364, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "ctime": 1734217400.0496364, 
        "dev": 5, 
        "device_type": 64768, 
        "executable": false, 
        "exists": true, 
        "gid": 6, 
        "gr_name": "disk", 
        "inode": 453486, 
        "isblk": true, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": false, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/symlink", 
        "mode": "0660", 
        "mtime": 1734217400.0496364, 
        "nlink": 1, 
        "path": "/dev/mapper/foo-test1", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": null, 
        "wgrp": true, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Saturday 14 December 2024  18:03:36 -0500 (0:00:00.489)       0:09:23.848 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Saturday 14 December 2024  18:03:36 -0500 (0:00:00.082)       0:09:23.930 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Saturday 14 December 2024  18:03:36 -0500 (0:00:00.069)       0:09:23.999 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Saturday 14 December 2024  18:03:36 -0500 (0:00:00.099)       0:09:24.099 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    }, 
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Saturday 14 December 2024  18:03:36 -0500 (0:00:00.093)       0:09:24.192 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Saturday 14 December 2024  18:03:36 -0500 (0:00:00.051)       0:09:24.244 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Saturday 14 December 2024  18:03:36 -0500 (0:00:00.067)       0:09:24.312 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217400.1626365, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "ctime": 1734217400.1626365, 
        "dev": 5, 
        "device_type": 64769, 
        "executable": false, 
        "exists": true, 
        "gid": 6, 
        "gr_name": "disk", 
        "inode": 464492, 
        "isblk": true, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": false, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/symlink", 
        "mode": "0660", 
        "mtime": 1734217400.1626365, 
        "nlink": 1, 
        "path": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": null, 
        "wgrp": true, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Saturday 14 December 2024  18:03:37 -0500 (0:00:00.369)       0:09:24.681 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}
lsrpackages: cryptsetup

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Saturday 14 December 2024  18:03:37 -0500 (0:00:00.677)       0:09:25.358 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cryptsetup", 
        "luksDump", 
        "/dev/mapper/foo-test1"
    ], 
    "delta": "0:00:00.025775", 
    "end": "2024-12-14 18:03:38.288590", 
    "rc": 0, 
    "start": "2024-12-14 18:03:38.262815"
}

STDOUT:

LUKS header information for /dev/mapper/foo-test1

Version:       	1
Cipher name:   	aes
Cipher mode:   	xts-plain64
Hash spec:     	sha256
Payload offset:	8192
MK bits:       	512
MK digest:     	a8 74 77 d1 48 6a 80 f3 51 c9 06 56 fc d5 e2 5e 9e d6 91 e9 
MK salt:       	be 87 d0 ba 27 b4 25 cd 92 ca 6f 0b 5b 19 67 ca 
               	97 6e fc 33 30 f8 45 2a 6e 7c 2a 81 91 ba 99 ef 
MK iterations: 	23239
UUID:          	4ce986c5-0285-435c-8ffc-f915139facba

Key Slot 0: ENABLED
	Iterations:         	371834
	Salt:               	35 e5 dc 06 23 e6 2c 5a 5a c2 0f ed fa 25 47 86 
	                      	85 09 df ab c7 52 27 de 72 70 33 2d 5c 3c 8c 15 
	Key material offset:	8
	AF stripes:            	4000
Key Slot 1: DISABLED
Key Slot 2: DISABLED
Key Slot 3: DISABLED
Key Slot 4: DISABLED
Key Slot 5: DISABLED
Key Slot 6: DISABLED
Key Slot 7: DISABLED

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Saturday 14 December 2024  18:03:38 -0500 (0:00:00.407)       0:09:25.766 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Saturday 14 December 2024  18:03:38 -0500 (0:00:00.101)       0:09:25.867 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Saturday 14 December 2024  18:03:38 -0500 (0:00:00.097)       0:09:25.964 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Saturday 14 December 2024  18:03:38 -0500 (0:00:00.146)       0:09:26.111 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Saturday 14 December 2024  18:03:38 -0500 (0:00:00.127)       0:09:26.238 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Saturday 14 December 2024  18:03:38 -0500 (0:00:00.075)       0:09:26.314 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Saturday 14 December 2024  18:03:38 -0500 (0:00:00.064)       0:09:26.379 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Saturday 14 December 2024  18:03:39 -0500 (0:00:00.062)       0:09:26.441 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [
            "luks-4ce986c5-0285-435c-8ffc-f915139facba /dev/mapper/foo-test1 -"
        ], 
        "_storage_test_expected_crypttab_entries": "1", 
        "_storage_test_expected_crypttab_key_file": "-"
    }, 
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Saturday 14 December 2024  18:03:39 -0500 (0:00:00.073)       0:09:26.515 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Saturday 14 December 2024  18:03:39 -0500 (0:00:00.073)       0:09:26.589 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Saturday 14 December 2024  18:03:39 -0500 (0:00:00.074)       0:09:26.663 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Saturday 14 December 2024  18:03:39 -0500 (0:00:00.080)       0:09:26.744 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Saturday 14 December 2024  18:03:39 -0500 (0:00:00.110)       0:09:26.854 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null, 
        "_storage_test_expected_crypttab_entries": null, 
        "_storage_test_expected_crypttab_key_file": null
    }, 
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Saturday 14 December 2024  18:03:39 -0500 (0:00:00.063)       0:09:26.918 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Saturday 14 December 2024  18:03:39 -0500 (0:00:00.058)       0:09:26.976 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Saturday 14 December 2024  18:03:39 -0500 (0:00:00.070)       0:09:27.046 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Saturday 14 December 2024  18:03:39 -0500 (0:00:00.062)       0:09:27.109 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Saturday 14 December 2024  18:03:39 -0500 (0:00:00.070)       0:09:27.180 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Saturday 14 December 2024  18:03:39 -0500 (0:00:00.064)       0:09:27.244 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Saturday 14 December 2024  18:03:39 -0500 (0:00:00.066)       0:09:27.311 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Saturday 14 December 2024  18:03:39 -0500 (0:00:00.052)       0:09:27.363 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Saturday 14 December 2024  18:03:40 -0500 (0:00:00.065)       0:09:27.429 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Saturday 14 December 2024  18:03:40 -0500 (0:00:00.068)       0:09:27.497 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Saturday 14 December 2024  18:03:40 -0500 (0:00:00.054)       0:09:27.551 ***** 
ok: [managed-node2] => {
    "bytes": 4294967296, 
    "changed": false, 
    "lvm": "4g", 
    "parted": "4GiB", 
    "size": "4 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Saturday 14 December 2024  18:03:40 -0500 (0:00:00.371)       0:09:27.922 ***** 
ok: [managed-node2] => {
    "bytes": 4294967296, 
    "changed": false, 
    "lvm": "4g", 
    "parted": "4GiB", 
    "size": "4 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Saturday 14 December 2024  18:03:40 -0500 (0:00:00.406)       0:09:28.329 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_expected_size": "4294967296"
    }, 
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Saturday 14 December 2024  18:03:41 -0500 (0:00:00.082)       0:09:28.411 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Saturday 14 December 2024  18:03:41 -0500 (0:00:00.072)       0:09:28.483 ***** 
ok: [managed-node2] => {
    "bytes": 10737418240, 
    "changed": false, 
    "lvm": "10g", 
    "parted": "10GiB", 
    "size": "10 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Saturday 14 December 2024  18:03:41 -0500 (0:00:00.550)       0:09:29.034 ***** 
skipping: [managed-node2] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Saturday 14 December 2024  18:03:41 -0500 (0:00:00.083)       0:09:29.117 ***** 
skipping: [managed-node2] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Saturday 14 December 2024  18:03:41 -0500 (0:00:00.054)       0:09:29.171 ***** 
skipping: [managed-node2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Saturday 14 December 2024  18:03:41 -0500 (0:00:00.049)       0:09:29.220 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Saturday 14 December 2024  18:03:41 -0500 (0:00:00.050)       0:09:29.271 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Saturday 14 December 2024  18:03:41 -0500 (0:00:00.052)       0:09:29.323 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Saturday 14 December 2024  18:03:41 -0500 (0:00:00.045)       0:09:29.369 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Saturday 14 December 2024  18:03:42 -0500 (0:00:00.044)       0:09:29.413 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Saturday 14 December 2024  18:03:42 -0500 (0:00:00.053)       0:09:29.467 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Saturday 14 December 2024  18:03:42 -0500 (0:00:00.058)       0:09:29.526 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Saturday 14 December 2024  18:03:42 -0500 (0:00:00.062)       0:09:29.588 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Saturday 14 December 2024  18:03:42 -0500 (0:00:00.094)       0:09:29.682 ***** 
skipping: [managed-node2] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Saturday 14 December 2024  18:03:42 -0500 (0:00:00.108)       0:09:29.791 ***** 
skipping: [managed-node2] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Saturday 14 December 2024  18:03:42 -0500 (0:00:00.087)       0:09:29.878 ***** 
skipping: [managed-node2] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Saturday 14 December 2024  18:03:42 -0500 (0:00:00.059)       0:09:29.937 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Saturday 14 December 2024  18:03:42 -0500 (0:00:00.087)       0:09:30.025 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Saturday 14 December 2024  18:03:42 -0500 (0:00:00.064)       0:09:30.089 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Saturday 14 December 2024  18:03:42 -0500 (0:00:00.075)       0:09:30.165 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Saturday 14 December 2024  18:03:42 -0500 (0:00:00.060)       0:09:30.226 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Saturday 14 December 2024  18:03:42 -0500 (0:00:00.062)       0:09:30.289 ***** 
ok: [managed-node2] => {
    "storage_test_actual_size": {
        "bytes": 4294967296, 
        "changed": false, 
        "failed": false, 
        "lvm": "4g", 
        "parted": "4GiB", 
        "size": "4 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Saturday 14 December 2024  18:03:42 -0500 (0:00:00.064)       0:09:30.353 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Saturday 14 December 2024  18:03:43 -0500 (0:00:00.048)       0:09:30.402 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Saturday 14 December 2024  18:03:43 -0500 (0:00:00.067)       0:09:30.469 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "lvs", 
        "--noheadings", 
        "--nameprefixes", 
        "--units=b", 
        "--nosuffix", 
        "--unquoted", 
        "-o", 
        "name,attr,cache_total_blocks,chunk_size,segtype", 
        "foo/test1"
    ], 
    "delta": "0:00:00.019840", 
    "end": "2024-12-14 18:03:43.462608", 
    "rc": 0, 
    "start": "2024-12-14 18:03:43.442768"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Saturday 14 December 2024  18:03:43 -0500 (0:00:00.483)       0:09:30.953 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    }, 
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Saturday 14 December 2024  18:03:43 -0500 (0:00:00.080)       0:09:31.034 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Saturday 14 December 2024  18:03:43 -0500 (0:00:00.087)       0:09:31.121 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Saturday 14 December 2024  18:03:43 -0500 (0:00:00.065)       0:09:31.187 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Saturday 14 December 2024  18:03:43 -0500 (0:00:00.157)       0:09:31.345 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Saturday 14 December 2024  18:03:44 -0500 (0:00:00.063)       0:09:31.408 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Saturday 14 December 2024  18:03:44 -0500 (0:00:00.062)       0:09:31.471 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    }, 
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Saturday 14 December 2024  18:03:44 -0500 (0:00:00.060)       0:09:31.531 ***** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Saturday 14 December 2024  18:03:44 -0500 (0:00:00.053)       0:09:31.584 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null, 
        "storage_test_crypttab": null, 
        "storage_test_fstab": null
    }, 
    "changed": false
}

TASK [Clean up] ****************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:510
Saturday 14 December 2024  18:03:44 -0500 (0:00:00.058)       0:09:31.643 ***** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Saturday 14 December 2024  18:03:44 -0500 (0:00:00.197)       0:09:31.840 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Saturday 14 December 2024  18:03:44 -0500 (0:00:00.149)       0:09:31.990 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Saturday 14 December 2024  18:03:44 -0500 (0:00:00.121)       0:09:32.112 ***** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F", 
            "ext3": "-F", 
            "ext4": "-F"
        }, 
        "blivet_package_list": [
            "python-enum34", 
            "python-blivet3", 
            "libblockdev-crypto", 
            "libblockdev-dm", 
            "libblockdev-lvm", 
            "libblockdev-mdraid", 
            "libblockdev-swap", 
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Saturday 14 December 2024  18:03:44 -0500 (0:00:00.183)       0:09:32.296 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Saturday 14 December 2024  18:03:44 -0500 (0:00:00.069)       0:09:32.365 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Saturday 14 December 2024  18:03:45 -0500 (0:00:00.080)       0:09:32.446 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Saturday 14 December 2024  18:03:45 -0500 (0:00:00.066)       0:09:32.512 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Saturday 14 December 2024  18:03:45 -0500 (0:00:00.058)       0:09:32.571 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Saturday 14 December 2024  18:03:45 -0500 (0:00:00.151)       0:09:32.722 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", 
        "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", 
        "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", 
        "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", 
        "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", 
        "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", 
        "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", 
        "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed"
    ]
}
lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Saturday 14 December 2024  18:03:46 -0500 (0:00:01.469)       0:09:34.191 ***** 
ok: [managed-node2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Saturday 14 December 2024  18:03:46 -0500 (0:00:00.067)       0:09:34.259 ***** 
ok: [managed-node2] => {
    "storage_volumes": [
        {
            "disks": [
                "sda"
            ], 
            "name": "foo", 
            "state": "absent", 
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Saturday 14 December 2024  18:03:46 -0500 (0:00:00.091)       0:09:34.350 ***** 
ok: [managed-node2] => {
    "actions": [], 
    "changed": false, 
    "crypts": [], 
    "leaves": [], 
    "mounts": [], 
    "packages": [], 
    "pools": [], 
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Saturday 14 December 2024  18:03:51 -0500 (0:00:04.211)       0:09:38.562 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node2

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Saturday 14 December 2024  18:03:51 -0500 (0:00:00.090)       0:09:38.653 ***** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Saturday 14 December 2024  18:03:51 -0500 (0:00:00.071)       0:09:38.724 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Saturday 14 December 2024  18:03:51 -0500 (0:00:00.062)       0:09:38.787 ***** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Saturday 14 December 2024  18:03:51 -0500 (0:00:00.062)       0:09:38.849 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed"
    ]
}
lsrpackages: kpartx

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Saturday 14 December 2024  18:03:52 -0500 (0:00:00.814)       0:09:39.663 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blivet.service": {
                "name": "blivet.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dm-event.service": {
                "name": "dm-event.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "dmraid-activation.service": {
                "name": "dmraid-activation.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdcheck_start.service": {
                "name": "mdcheck_start.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmon@.service": {
                "name": "mdmon@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "mdmonitor.service": {
                "name": "mdmonitor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "target.service": {
                "name": "target.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "targetclid.service": {
                "name": "targetclid.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Saturday 14 December 2024  18:03:53 -0500 (0:00:01.465)       0:09:41.129 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Saturday 14 December 2024  18:03:53 -0500 (0:00:00.089)       0:09:41.218 ***** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Saturday 14 December 2024  18:03:53 -0500 (0:00:00.057)       0:09:41.276 ***** 
changed: [managed-node2] => {
    "actions": [
        {
            "action": "destroy format", 
            "device": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
            "fs_type": "xfs"
        }, 
        {
            "action": "destroy device", 
            "device": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
            "fs_type": null
        }, 
        {
            "action": "destroy format", 
            "device": "/dev/mapper/foo-test1", 
            "fs_type": "luks"
        }, 
        {
            "action": "destroy device", 
            "device": "/dev/mapper/foo-test1", 
            "fs_type": null
        }, 
        {
            "action": "destroy device", 
            "device": "/dev/foo", 
            "fs_type": null
        }, 
        {
            "action": "destroy format", 
            "device": "/dev/sda", 
            "fs_type": "lvmpv"
        }
    ], 
    "changed": true, 
    "crypts": [
        {
            "backing_device": "/dev/mapper/foo-test1", 
            "name": "luks-4ce986c5-0285-435c-8ffc-f915139facba", 
            "password": "-", 
            "state": "absent"
        }
    ], 
    "leaves": [
        "/dev/sda", 
        "/dev/sdb", 
        "/dev/sdc", 
        "/dev/sdd", 
        "/dev/sde", 
        "/dev/sdf", 
        "/dev/sdg", 
        "/dev/sdh", 
        "/dev/sdi", 
        "/dev/xvda1"
    ], 
    "mounts": [
        {
            "fstype": "xfs", 
            "path": "/opt/test1", 
            "src": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
            "state": "absent"
        }
    ], 
    "packages": [
        "e2fsprogs"
    ], 
    "pools": [], 
    "volumes": [
        {
            "_device": "/dev/sda", 
            "_mount_id": "UUID=xa7q16-X4RQ-vFWY-Ftcj-AFkA-0I7B-z0LDg5", 
            "_raw_device": "/dev/sda", 
            "cache_devices": [], 
            "cache_mode": null, 
            "cache_size": 0, 
            "cached": false, 
            "compression": null, 
            "deduplication": null, 
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_cipher": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": null, 
            "fs_create_options": "", 
            "fs_label": "", 
            "fs_overwrite_existing": true, 
            "fs_type": "lvmpv", 
            "mount_check": 0, 
            "mount_device_identifier": "uuid", 
            "mount_group": null, 
            "mount_mode": null, 
            "mount_options": "defaults", 
            "mount_passno": 0, 
            "mount_point": null, 
            "mount_user": null, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "raid_stripe_size": null, 
            "size": 10737418240, 
            "state": "absent", 
            "thin": null, 
            "thin_pool_name": null, 
            "thin_pool_size": null, 
            "type": "disk", 
            "vdo_pool_size": null
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Saturday 14 December 2024  18:04:28 -0500 (0:00:35.005)       0:10:16.281 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Saturday 14 December 2024  18:04:28 -0500 (0:00:00.059)       0:10:16.340 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217403.030635, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "4d3e07560ed61cbf7ba7ae4fc788fd959ef2ec87", 
        "ctime": 1734217403.027635, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 263588, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0644", 
        "mtime": 1734217403.027635, 
        "nlink": 1, 
        "path": "/etc/fstab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 1311, 
        "uid": 0, 
        "version": "606402765", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Saturday 14 December 2024  18:04:29 -0500 (0:00:00.486)       0:10:16.827 ***** 
ok: [managed-node2] => {
    "backup": "", 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Saturday 14 December 2024  18:04:30 -0500 (0:00:00.600)       0:10:17.428 ***** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Saturday 14 December 2024  18:04:30 -0500 (0:00:00.052)       0:10:17.480 ***** 
ok: [managed-node2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format", 
                "device": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
                "fs_type": "xfs"
            }, 
            {
                "action": "destroy device", 
                "device": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
                "fs_type": null
            }, 
            {
                "action": "destroy format", 
                "device": "/dev/mapper/foo-test1", 
                "fs_type": "luks"
            }, 
            {
                "action": "destroy device", 
                "device": "/dev/mapper/foo-test1", 
                "fs_type": null
            }, 
            {
                "action": "destroy device", 
                "device": "/dev/foo", 
                "fs_type": null
            }, 
            {
                "action": "destroy format", 
                "device": "/dev/sda", 
                "fs_type": "lvmpv"
            }
        ], 
        "changed": true, 
        "crypts": [
            {
                "backing_device": "/dev/mapper/foo-test1", 
                "name": "luks-4ce986c5-0285-435c-8ffc-f915139facba", 
                "password": "-", 
                "state": "absent"
            }
        ], 
        "failed": false, 
        "leaves": [
            "/dev/sda", 
            "/dev/sdb", 
            "/dev/sdc", 
            "/dev/sdd", 
            "/dev/sde", 
            "/dev/sdf", 
            "/dev/sdg", 
            "/dev/sdh", 
            "/dev/sdi", 
            "/dev/xvda1"
        ], 
        "mounts": [
            {
                "fstype": "xfs", 
                "path": "/opt/test1", 
                "src": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
                "state": "absent"
            }
        ], 
        "packages": [
            "e2fsprogs"
        ], 
        "pools": [], 
        "volumes": [
            {
                "_device": "/dev/sda", 
                "_mount_id": "UUID=xa7q16-X4RQ-vFWY-Ftcj-AFkA-0I7B-z0LDg5", 
                "_raw_device": "/dev/sda", 
                "cache_devices": [], 
                "cache_mode": null, 
                "cache_size": 0, 
                "cached": false, 
                "compression": null, 
                "deduplication": null, 
                "disks": [
                    "sda"
                ], 
                "encryption": false, 
                "encryption_cipher": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": null, 
                "fs_create_options": "", 
                "fs_label": "", 
                "fs_overwrite_existing": true, 
                "fs_type": "lvmpv", 
                "mount_check": 0, 
                "mount_device_identifier": "uuid", 
                "mount_group": null, 
                "mount_mode": null, 
                "mount_options": "defaults", 
                "mount_passno": 0, 
                "mount_point": null, 
                "mount_user": null, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "raid_stripe_size": null, 
                "size": 10737418240, 
                "state": "absent", 
                "thin": null, 
                "thin_pool_name": null, 
                "thin_pool_size": null, 
                "type": "disk", 
                "vdo_pool_size": null
            }
        ]
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Saturday 14 December 2024  18:04:30 -0500 (0:00:00.081)       0:10:17.561 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Saturday 14 December 2024  18:04:30 -0500 (0:00:00.064)       0:10:17.626 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_volumes_list": [
            {
                "_device": "/dev/sda", 
                "_mount_id": "UUID=xa7q16-X4RQ-vFWY-Ftcj-AFkA-0I7B-z0LDg5", 
                "_raw_device": "/dev/sda", 
                "cache_devices": [], 
                "cache_mode": null, 
                "cache_size": 0, 
                "cached": false, 
                "compression": null, 
                "deduplication": null, 
                "disks": [
                    "sda"
                ], 
                "encryption": false, 
                "encryption_cipher": null, 
                "encryption_key": null, 
                "encryption_key_size": null, 
                "encryption_luks_version": null, 
                "encryption_password": null, 
                "fs_create_options": "", 
                "fs_label": "", 
                "fs_overwrite_existing": true, 
                "fs_type": "lvmpv", 
                "mount_check": 0, 
                "mount_device_identifier": "uuid", 
                "mount_group": null, 
                "mount_mode": null, 
                "mount_options": "defaults", 
                "mount_passno": 0, 
                "mount_point": null, 
                "mount_user": null, 
                "name": "foo", 
                "raid_chunk_size": null, 
                "raid_device_count": null, 
                "raid_level": null, 
                "raid_metadata_version": null, 
                "raid_spare_count": null, 
                "raid_stripe_size": null, 
                "size": 10737418240, 
                "state": "absent", 
                "thin": null, 
                "thin_pool_name": null, 
                "thin_pool_size": null, 
                "type": "disk", 
                "vdo_pool_size": null
            }
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Saturday 14 December 2024  18:04:30 -0500 (0:00:00.091)       0:10:17.717 ***** 
changed: [managed-node2] => (item={u'src': u'/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba', u'state': u'absent', u'fstype': u'xfs', u'path': u'/opt/test1'}) => {
    "ansible_loop_var": "mount_info", 
    "changed": true, 
    "dump": "0", 
    "fstab": "/etc/fstab", 
    "fstype": "xfs", 
    "mount_info": {
        "fstype": "xfs", 
        "path": "/opt/test1", 
        "src": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba", 
        "state": "absent"
    }, 
    "name": "/opt/test1", 
    "opts": "defaults", 
    "passno": "0", 
    "src": "/dev/mapper/luks-4ce986c5-0285-435c-8ffc-f915139facba"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Saturday 14 December 2024  18:04:30 -0500 (0:00:00.584)       0:10:18.302 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Saturday 14 December 2024  18:04:31 -0500 (0:00:00.749)       0:10:19.051 ***** 

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Saturday 14 December 2024  18:04:31 -0500 (0:00:00.124)       0:10:19.175 ***** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Saturday 14 December 2024  18:04:31 -0500 (0:00:00.090)       0:10:19.266 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "name": null, 
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Saturday 14 December 2024  18:04:32 -0500 (0:00:00.796)       0:10:20.062 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217407.5336332, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "8e3b36fb89530f0cb78748f8af3193ebdc43ed2e", 
        "ctime": 1734217404.7966344, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 917511, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0600", 
        "mtime": 1734217404.7956343, 
        "nlink": 1, 
        "path": "/etc/crypttab", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": false, 
        "roth": false, 
        "rusr": true, 
        "size": 66, 
        "uid": 0, 
        "version": "606410125", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Saturday 14 December 2024  18:04:33 -0500 (0:00:00.522)       0:10:20.585 ***** 
changed: [managed-node2] => (item={u'state': u'absent', u'password': u'-', u'name': u'luks-4ce986c5-0285-435c-8ffc-f915139facba', u'backing_device': u'/dev/mapper/foo-test1'}) => {
    "ansible_loop_var": "entry", 
    "backup": "", 
    "changed": true, 
    "entry": {
        "backing_device": "/dev/mapper/foo-test1", 
        "name": "luks-4ce986c5-0285-435c-8ffc-f915139facba", 
        "password": "-", 
        "state": "absent"
    }, 
    "found": 1
}

MSG:

1 line(s) removed

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Saturday 14 December 2024  18:04:33 -0500 (0:00:00.388)       0:10:20.973 ***** 
ok: [managed-node2]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks.yml:520
Saturday 14 December 2024  18:04:34 -0500 (0:00:00.755)       0:10:21.728 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node2

TASK [Print out pool information] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Saturday 14 December 2024  18:04:34 -0500 (0:00:00.095)       0:10:21.823 ***** 
skipping: [managed-node2] => {}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Saturday 14 December 2024  18:04:34 -0500 (0:00:00.038)       0:10:21.862 ***** 
ok: [managed-node2] => {
    "_storage_volumes_list": [
        {
            "_device": "/dev/sda", 
            "_mount_id": "UUID=xa7q16-X4RQ-vFWY-Ftcj-AFkA-0I7B-z0LDg5", 
            "_raw_device": "/dev/sda", 
            "cache_devices": [], 
            "cache_mode": null, 
            "cache_size": 0, 
            "cached": false, 
            "compression": null, 
            "deduplication": null, 
            "disks": [
                "sda"
            ], 
            "encryption": false, 
            "encryption_cipher": null, 
            "encryption_key": null, 
            "encryption_key_size": null, 
            "encryption_luks_version": null, 
            "encryption_password": null, 
            "fs_create_options": "", 
            "fs_label": "", 
            "fs_overwrite_existing": true, 
            "fs_type": "lvmpv", 
            "mount_check": 0, 
            "mount_device_identifier": "uuid", 
            "mount_group": null, 
            "mount_mode": null, 
            "mount_options": "defaults", 
            "mount_passno": 0, 
            "mount_point": null, 
            "mount_user": null, 
            "name": "foo", 
            "raid_chunk_size": null, 
            "raid_device_count": null, 
            "raid_level": null, 
            "raid_metadata_version": null, 
            "raid_spare_count": null, 
            "raid_stripe_size": null, 
            "size": 10737418240, 
            "state": "absent", 
            "thin": null, 
            "thin_pool_name": null, 
            "thin_pool_size": null, 
            "type": "disk", 
            "vdo_pool_size": null
        }
    ]
}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Saturday 14 December 2024  18:04:34 -0500 (0:00:00.046)       0:10:21.909 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "info": {
        "/dev/sda": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sda", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdb": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdb", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdc": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdc", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdd": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdd", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sde": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sde", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdf": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdf", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdg": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdg", 
            "size": "1T", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdh": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdh", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/sdi": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/sdi", 
            "size": "10G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda": {
            "fstype": "", 
            "label": "", 
            "mountpoint": "", 
            "name": "/dev/xvda", 
            "size": "250G", 
            "type": "disk", 
            "uuid": ""
        }, 
        "/dev/xvda1": {
            "fstype": "ext4", 
            "label": "", 
            "mountpoint": "/", 
            "name": "/dev/xvda1", 
            "size": "250G", 
            "type": "partition", 
            "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Saturday 14 December 2024  18:04:34 -0500 (0:00:00.358)       0:10:22.268 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/fstab"
    ], 
    "delta": "0:00:00.002909", 
    "end": "2024-12-14 18:04:35.190487", 
    "rc": 0, 
    "start": "2024-12-14 18:04:35.187578"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Thu Jun 20 10:23:46 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 /                       ext4    defaults        1 1
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Saturday 14 December 2024  18:04:35 -0500 (0:00:00.393)       0:10:22.661 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "cat", 
        "/etc/crypttab"
    ], 
    "delta": "0:00:00.003524", 
    "end": "2024-12-14 18:04:35.594258", 
    "failed_when_result": false, 
    "rc": 0, 
    "start": "2024-12-14 18:04:35.590734"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Saturday 14 December 2024  18:04:35 -0500 (0:00:00.401)       0:10:23.063 ***** 

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Saturday 14 December 2024  18:04:35 -0500 (0:00:00.054)       0:10:23.117 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node2

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Saturday 14 December 2024  18:04:35 -0500 (0:00:00.116)       0:10:23.234 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": false, 
        "_storage_volume_tests": [
            "mount", 
            "fstab", 
            "fs", 
            "device", 
            "encryption", 
            "md", 
            "size", 
            "cache"
        ]
    }, 
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Saturday 14 December 2024  18:04:35 -0500 (0:00:00.069)       0:10:23.303 ***** 
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node2
included: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node2

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Saturday 14 December 2024  18:04:36 -0500 (0:00:00.269)       0:10:23.572 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/sda"
    }, 
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Saturday 14 December 2024  18:04:36 -0500 (0:00:00.081)       0:10:23.654 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "", 
        "storage_test_swap_expected_matches": "0"
    }, 
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Saturday 14 December 2024  18:04:36 -0500 (0:00:00.163)       0:10:23.818 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Saturday 14 December 2024  18:04:36 -0500 (0:00:00.056)       0:10:23.874 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Saturday 14 December 2024  18:04:36 -0500 (0:00:00.052)       0:10:23.927 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Saturday 14 December 2024  18:04:36 -0500 (0:00:00.056)       0:10:23.983 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Saturday 14 December 2024  18:04:36 -0500 (0:00:00.055)       0:10:24.039 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Saturday 14 December 2024  18:04:36 -0500 (0:00:00.056)       0:10:24.096 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Saturday 14 December 2024  18:04:36 -0500 (0:00:00.060)       0:10:24.156 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Saturday 14 December 2024  18:04:36 -0500 (0:00:00.057)       0:10:24.214 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Saturday 14 December 2024  18:04:36 -0500 (0:00:00.056)       0:10:24.271 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null, 
        "storage_test_mount_expected_mount_point": null, 
        "storage_test_swap_expected_matches": null, 
        "storage_test_swaps": null, 
        "storage_test_sys_node": null
    }, 
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Saturday 14 December 2024  18:04:36 -0500 (0:00:00.057)       0:10:24.329 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "0", 
        "storage_test_fstab_expected_mount_options_matches": "0", 
        "storage_test_fstab_expected_mount_point_matches": "0", 
        "storage_test_fstab_id_matches": [], 
        "storage_test_fstab_mount_options_matches": [], 
        "storage_test_fstab_mount_point_matches": []
    }, 
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Saturday 14 December 2024  18:04:37 -0500 (0:00:00.095)       0:10:24.424 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Saturday 14 December 2024  18:04:37 -0500 (0:00:00.055)       0:10:24.479 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Saturday 14 December 2024  18:04:37 -0500 (0:00:00.071)       0:10:24.550 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Saturday 14 December 2024  18:04:37 -0500 (0:00:00.056)       0:10:24.607 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Saturday 14 December 2024  18:04:37 -0500 (0:00:00.116)       0:10:24.724 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null, 
        "storage_test_fstab_expected_mount_options_matches": null, 
        "storage_test_fstab_expected_mount_point_matches": null, 
        "storage_test_fstab_id_matches": null, 
        "storage_test_fstab_mount_options_matches": null, 
        "storage_test_fstab_mount_point_matches": null
    }, 
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Saturday 14 December 2024  18:04:37 -0500 (0:00:00.098)       0:10:24.822 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Saturday 14 December 2024  18:04:37 -0500 (0:00:00.099)       0:10:24.922 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Saturday 14 December 2024  18:04:37 -0500 (0:00:00.072)       0:10:24.995 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1734217468.7166057, 
        "attr_flags": "", 
        "attributes": [], 
        "block_size": 4096, 
        "blocks": 0, 
        "charset": "binary", 
        "ctime": 1734217468.7166057, 
        "dev": 5, 
        "device_type": 2048, 
        "executable": false, 
        "exists": true, 
        "gid": 6, 
        "gr_name": "disk", 
        "inode": 28408, 
        "isblk": true, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": false, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "inode/blockdevice", 
        "mode": "0660", 
        "mtime": 1734217468.7166057, 
        "nlink": 1, 
        "path": "/dev/sda", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": false, 
        "rusr": true, 
        "size": 0, 
        "uid": 0, 
        "version": null, 
        "wgrp": true, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Saturday 14 December 2024  18:04:38 -0500 (0:00:00.439)       0:10:25.435 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Saturday 14 December 2024  18:04:38 -0500 (0:00:00.078)       0:10:25.513 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Saturday 14 December 2024  18:04:38 -0500 (0:00:00.056)       0:10:25.570 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Saturday 14 December 2024  18:04:38 -0500 (0:00:00.062)       0:10:25.633 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "st_volume_type": "disk"
    }, 
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Saturday 14 December 2024  18:04:38 -0500 (0:00:00.079)       0:10:25.712 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Saturday 14 December 2024  18:04:38 -0500 (0:00:00.049)       0:10:25.762 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Saturday 14 December 2024  18:04:38 -0500 (0:00:00.056)       0:10:25.819 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Saturday 14 December 2024  18:04:38 -0500 (0:00:00.042)       0:10:25.861 ***** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}
lsrpackages: cryptsetup

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Saturday 14 December 2024  18:04:39 -0500 (0:00:00.881)       0:10:26.742 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Saturday 14 December 2024  18:04:39 -0500 (0:00:00.101)       0:10:26.843 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Saturday 14 December 2024  18:04:39 -0500 (0:00:00.087)       0:10:26.931 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Saturday 14 December 2024  18:04:39 -0500 (0:00:00.053)       0:10:26.985 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Saturday 14 December 2024  18:04:39 -0500 (0:00:00.056)       0:10:27.042 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Saturday 14 December 2024  18:04:39 -0500 (0:00:00.065)       0:10:27.107 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Saturday 14 December 2024  18:04:39 -0500 (0:00:00.039)       0:10:27.147 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Saturday 14 December 2024  18:04:39 -0500 (0:00:00.049)       0:10:27.196 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Saturday 14 December 2024  18:04:39 -0500 (0:00:00.039)       0:10:27.236 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [], 
        "_storage_test_expected_crypttab_entries": "0", 
        "_storage_test_expected_crypttab_key_file": "-"
    }, 
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Saturday 14 December 2024  18:04:39 -0500 (0:00:00.053)       0:10:27.289 ***** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Saturday 14 December 2024  18:04:39 -0500 (0:00:00.043)       0:10:27.333 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Saturday 14 December 2024  18:04:39 -0500 (0:00:00.045)       0:10:27.378 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Saturday 14 December 2024  18:04:40 -0500 (0:00:00.057)       0:10:27.436 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Saturday 14 December 2024  18:04:40 -0500 (0:00:00.098)       0:10:27.535 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null, 
        "_storage_test_expected_crypttab_entries": null, 
        "_storage_test_expected_crypttab_key_file": null
    }, 
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Saturday 14 December 2024  18:04:40 -0500 (0:00:00.059)       0:10:27.594 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Saturday 14 December 2024  18:04:40 -0500 (0:00:00.081)       0:10:27.675 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Saturday 14 December 2024  18:04:40 -0500 (0:00:00.104)       0:10:27.780 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Saturday 14 December 2024  18:04:40 -0500 (0:00:00.068)       0:10:27.849 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Saturday 14 December 2024  18:04:40 -0500 (0:00:00.066)       0:10:27.916 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Saturday 14 December 2024  18:04:40 -0500 (0:00:00.078)       0:10:27.995 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Saturday 14 December 2024  18:04:40 -0500 (0:00:00.056)       0:10:28.051 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Saturday 14 December 2024  18:04:40 -0500 (0:00:00.056)       0:10:28.108 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Saturday 14 December 2024  18:04:40 -0500 (0:00:00.074)       0:10:28.182 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Saturday 14 December 2024  18:04:40 -0500 (0:00:00.101)       0:10:28.284 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Saturday 14 December 2024  18:04:40 -0500 (0:00:00.097)       0:10:28.381 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Saturday 14 December 2024  18:04:41 -0500 (0:00:00.061)       0:10:28.443 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Saturday 14 December 2024  18:04:41 -0500 (0:00:00.059)       0:10:28.502 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Saturday 14 December 2024  18:04:41 -0500 (0:00:00.080)       0:10:28.583 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Saturday 14 December 2024  18:04:41 -0500 (0:00:00.071)       0:10:28.654 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Saturday 14 December 2024  18:04:41 -0500 (0:00:00.074)       0:10:28.729 ***** 
skipping: [managed-node2] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Saturday 14 December 2024  18:04:41 -0500 (0:00:00.056)       0:10:28.786 ***** 
skipping: [managed-node2] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Saturday 14 December 2024  18:04:41 -0500 (0:00:00.052)       0:10:28.839 ***** 
skipping: [managed-node2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Saturday 14 December 2024  18:04:41 -0500 (0:00:00.055)       0:10:28.895 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Saturday 14 December 2024  18:04:41 -0500 (0:00:00.060)       0:10:28.955 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Saturday 14 December 2024  18:04:41 -0500 (0:00:00.104)       0:10:29.059 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Saturday 14 December 2024  18:04:41 -0500 (0:00:00.064)       0:10:29.123 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Saturday 14 December 2024  18:04:41 -0500 (0:00:00.065)       0:10:29.189 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Saturday 14 December 2024  18:04:41 -0500 (0:00:00.066)       0:10:29.255 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Saturday 14 December 2024  18:04:41 -0500 (0:00:00.083)       0:10:29.338 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Saturday 14 December 2024  18:04:42 -0500 (0:00:00.073)       0:10:29.412 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Saturday 14 December 2024  18:04:42 -0500 (0:00:00.056)       0:10:29.469 ***** 
skipping: [managed-node2] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Saturday 14 December 2024  18:04:42 -0500 (0:00:00.076)       0:10:29.545 ***** 
skipping: [managed-node2] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Saturday 14 December 2024  18:04:42 -0500 (0:00:00.169)       0:10:29.715 ***** 
skipping: [managed-node2] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Saturday 14 December 2024  18:04:42 -0500 (0:00:00.065)       0:10:29.781 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Saturday 14 December 2024  18:04:42 -0500 (0:00:00.069)       0:10:29.850 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Saturday 14 December 2024  18:04:42 -0500 (0:00:00.062)       0:10:29.913 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Saturday 14 December 2024  18:04:42 -0500 (0:00:00.069)       0:10:29.983 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Saturday 14 December 2024  18:04:42 -0500 (0:00:00.067)       0:10:30.050 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Saturday 14 December 2024  18:04:42 -0500 (0:00:00.069)       0:10:30.120 ***** 
ok: [managed-node2] => {
    "storage_test_actual_size": {
        "changed": false, 
        "skip_reason": "Conditional result was False", 
        "skipped": true
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Saturday 14 December 2024  18:04:42 -0500 (0:00:00.076)       0:10:30.197 ***** 
ok: [managed-node2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Saturday 14 December 2024  18:04:42 -0500 (0:00:00.062)       0:10:30.259 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Saturday 14 December 2024  18:04:42 -0500 (0:00:00.055)       0:10:30.314 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Saturday 14 December 2024  18:04:42 -0500 (0:00:00.055)       0:10:30.370 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Saturday 14 December 2024  18:04:43 -0500 (0:00:00.057)       0:10:30.427 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Saturday 14 December 2024  18:04:43 -0500 (0:00:00.059)       0:10:30.487 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Saturday 14 December 2024  18:04:43 -0500 (0:00:00.060)       0:10:30.547 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Saturday 14 December 2024  18:04:43 -0500 (0:00:00.073)       0:10:30.621 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Saturday 14 December 2024  18:04:43 -0500 (0:00:00.058)       0:10:30.680 ***** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Saturday 14 December 2024  18:04:43 -0500 (0:00:00.069)       0:10:30.749 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    }, 
    "changed": false
}

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Saturday 14 December 2024  18:04:43 -0500 (0:00:00.117)       0:10:30.867 ***** 
ok: [managed-node2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null, 
        "storage_test_crypttab": null, 
        "storage_test_fstab": null
    }, 
    "changed": false
}
META: ran handlers
META: ran handlers

PLAY RECAP *********************************************************************
managed-node2              : ok=1224 changed=60   unreachable=0    failed=9    skipped=1066 rescued=9    ignored=0   

Saturday 14 December 2024  18:04:43 -0500 (0:00:00.105)       0:10:30.972 ***** 
=============================================================================== 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state -- 65.27s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state -- 35.01s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state -- 11.02s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state -- 10.88s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state -- 10.77s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state -- 10.56s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state -- 10.36s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state -- 10.35s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Get required packages --------------- 4.49s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 
fedora.linux_system_roles.storage : Get required packages --------------- 4.40s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.38s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.36s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.35s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.33s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.27s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Get required packages --------------- 4.26s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 
fedora.linux_system_roles.storage : Get required packages --------------- 4.21s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 
fedora.linux_system_roles.storage : Get required packages --------------- 4.21s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 
fedora.linux_system_roles.storage : Get required packages --------------- 4.20s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 
fedora.linux_system_roles.storage : Get required packages --------------- 4.20s
/tmp/collections-D2m/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19