ansible-playbook 2.9.27
  config file = None
  configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
  ansible python module location = /usr/local/lib/python3.9/site-packages/ansible
  executable location = /usr/local/bin/ansible-playbook
  python version = 3.9.19 (main, May 16 2024, 11:40:09) [GCC 8.5.0 20210514 (Red Hat 8.5.0-22)]
No config file found; using defaults
[WARNING]: running playbook inside collection fedora.linux_system_roles
statically imported: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml
statically imported: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml
statically imported: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml
statically imported: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml
statically imported: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml
statically imported: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml
statically imported: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml
statically imported: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml
statically imported: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml
statically imported: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml
statically imported: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml
statically imported: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml
Skipping callback 'actionable', as we already have a stdout callback.
Skipping callback 'counter_enabled', as we already have a stdout callback.
Skipping callback 'debug', as we already have a stdout callback.
Skipping callback 'dense', as we already have a stdout callback.
Skipping callback 'dense', as we already have a stdout callback.
Skipping callback 'full_skip', as we already have a stdout callback.
Skipping callback 'json', as we already have a stdout callback.
Skipping callback 'minimal', as we already have a stdout callback.
Skipping callback 'null', as we already have a stdout callback.
Skipping callback 'oneline', as we already have a stdout callback.
Skipping callback 'selective', as we already have a stdout callback.
Skipping callback 'skippy', as we already have a stdout callback.
Skipping callback 'stderr', as we already have a stdout callback.
Skipping callback 'unixy', as we already have a stdout callback.
Skipping callback 'yaml', as we already have a stdout callback.

PLAYBOOK: tests_luks2.yml ******************************************************
1 plays in /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml

PLAY [Test LUKS2] **************************************************************

TASK [Gathering Facts] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:2
Monday 20 January 2025  06:11:31 -0500 (0:00:00.023)       0:00:00.023 ******** 
ok: [managed-node1]
META: ran handlers

TASK [Enable FIPS mode] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:20
Monday 20 January 2025  06:11:32 -0500 (0:00:01.019)       0:00:01.042 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reboot] ******************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:28
Monday 20 January 2025  06:11:32 -0500 (0:00:00.037)       0:00:01.079 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Enable FIPS mode] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:39
Monday 20 January 2025  06:11:32 -0500 (0:00:00.037)       0:00:01.117 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reboot] ******************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:43
Monday 20 January 2025  06:11:32 -0500 (0:00:00.036)       0:00:01.154 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure dracut-fips] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:53
Monday 20 January 2025  06:11:32 -0500 (0:00:00.056)       0:00:01.210 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Configure boot for FIPS] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:59
Monday 20 January 2025  06:11:32 -0500 (0:00:00.071)       0:00:01.281 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reboot] ******************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:68
Monday 20 January 2025  06:11:32 -0500 (0:00:00.038)       0:00:01.320 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Run the role] ************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:72
Monday 20 January 2025  06:11:32 -0500 (0:00:00.036)       0:00:01.356 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:11:32 -0500 (0:00:00.033)       0:00:01.389 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:11:32 -0500 (0:00:00.025)       0:00:01.415 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:11:32 -0500 (0:00:00.037)       0:00:01.453 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:11:32 -0500 (0:00:00.064)       0:00:01.517 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "exists": false
    }
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:11:33 -0500 (0:00:00.444)       0:00:01.962 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "__storage_is_ostree": false
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:11:33 -0500 (0:00:00.044)       0:00:02.007 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:11:33 -0500 (0:00:00.021)       0:00:02.028 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:11:33 -0500 (0:00:00.020)       0:00:02.049 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:11:33 -0500 (0:00:00.091)       0:00:02.141 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:11:36 -0500 (0:00:02.971)       0:00:05.112 ******** 
ok: [managed-node1] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:11:36 -0500 (0:00:00.037)       0:00:05.149 ******** 
ok: [managed-node1] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:11:36 -0500 (0:00:00.037)       0:00:05.187 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:11:37 -0500 (0:00:00.650)       0:00:05.837 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:11:37 -0500 (0:00:00.065)       0:00:05.903 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:11:37 -0500 (0:00:00.014)       0:00:05.918 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:11:37 -0500 (0:00:00.016)       0:00:05.934 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:11:37 -0500 (0:00:00.014)       0:00:05.948 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:11:39 -0500 (0:00:02.802)       0:00:08.750 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:11:41 -0500 (0:00:01.716)       0:00:10.466 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:11:41 -0500 (0:00:00.032)       0:00:10.499 ******** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:11:41 -0500 (0:00:00.022)       0:00:10.521 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Monday 20 January 2025  06:11:42 -0500 (0:00:00.525)       0:00:11.047 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Monday 20 January 2025  06:11:42 -0500 (0:00:00.020)       0:00:11.067 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371482.0043535,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "ab8070345774adad92683e9645714452be7be474",
        "ctime": 1737371481.4993525,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 207618244,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1737371481.4993525,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1343,
        "uid": 0,
        "version": "261612921",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Monday 20 January 2025  06:11:42 -0500 (0:00:00.358)       0:00:11.425 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:11:42 -0500 (0:00:00.031)       0:00:11.457 ******** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Monday 20 January 2025  06:11:42 -0500 (0:00:00.031)       0:00:11.488 ******** 
ok: [managed-node1] => {
    "blivet_output": {
        "actions": [],
        "changed": false,
        "crypts": [],
        "failed": false,
        "leaves": [],
        "mounts": [],
        "packages": [],
        "pools": [],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Monday 20 January 2025  06:11:42 -0500 (0:00:00.035)       0:00:11.524 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Monday 20 January 2025  06:11:42 -0500 (0:00:00.040)       0:00:11.564 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Monday 20 January 2025  06:11:42 -0500 (0:00:00.029)       0:00:11.594 ******** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Monday 20 January 2025  06:11:42 -0500 (0:00:00.032)       0:00:11.626 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Monday 20 January 2025  06:11:42 -0500 (0:00:00.027)       0:00:11.654 ******** 

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Monday 20 January 2025  06:11:42 -0500 (0:00:00.023)       0:00:11.678 ******** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Monday 20 January 2025  06:11:42 -0500 (0:00:00.019)       0:00:11.697 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Monday 20 January 2025  06:11:42 -0500 (0:00:00.020)       0:00:11.718 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371407.2601843,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1737371405.34018,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 283116674,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1737371405.33918,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1693684562",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Monday 20 January 2025  06:11:43 -0500 (0:00:00.372)       0:00:12.090 ******** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Monday 20 January 2025  06:11:43 -0500 (0:00:00.017)       0:00:12.108 ******** 
ok: [managed-node1]

TASK [Get unused disks] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:76
Monday 20 January 2025  06:11:44 -0500 (0:00:00.754)       0:00:12.863 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml for managed-node1

TASK [Ensure test packages] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:2
Monday 20 January 2025  06:11:44 -0500 (0:00:00.033)       0:00:12.896 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Find unused disks in the system] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:11
Monday 20 January 2025  06:11:46 -0500 (0:00:02.864)       0:00:15.761 ******** 
ok: [managed-node1] => {
    "changed": false,
    "disks": [
        "sda"
    ],
    "info": [
        "Line: NAME=\"/dev/sda\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/sdb\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/sdc\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/sdd\" TYPE=\"disk\" SIZE=\"1099511627776\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/sde\" TYPE=\"disk\" SIZE=\"1099511627776\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/sdf\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/sdg\" TYPE=\"disk\" SIZE=\"1099511627776\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/sdh\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/sdi\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/xvda\" TYPE=\"disk\" SIZE=\"268435456000\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/xvda1\" TYPE=\"part\" SIZE=\"268434390528\" FSTYPE=\"xfs\" LOG-SEC=\"512\"",
        "Line type [part] is not disk: NAME=\"/dev/xvda1\" TYPE=\"part\" SIZE=\"268434390528\" FSTYPE=\"xfs\" LOG-SEC=\"512\"",
        "filename [xvda1] is a partition",
        "Disk [/dev/xvda] attrs [{'type': 'disk', 'size': '268435456000', 'fstype': '', 'ssize': '512'}] has partitions"
    ]
}

TASK [Debug why there are no unused disks] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:20
Monday 20 January 2025  06:11:48 -0500 (0:00:01.608)       0:00:17.369 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set unused_disks if necessary] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:29
Monday 20 January 2025  06:11:48 -0500 (0:00:00.036)       0:00:17.405 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "unused_disks": [
            "sda"
        ]
    },
    "changed": false
}

TASK [Exit playbook when there's not enough unused disks in the system] ********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:34
Monday 20 January 2025  06:11:48 -0500 (0:00:00.067)       0:00:17.473 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Print unused disks] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:39
Monday 20 January 2025  06:11:48 -0500 (0:00:00.030)       0:00:17.503 ******** 
ok: [managed-node1] => {
    "unused_disks": [
        "sda"
    ]
}

TASK [Test for correct handling of new encrypted volume w/ no key] *************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:85
Monday 20 January 2025  06:11:48 -0500 (0:00:00.031)       0:00:17.535 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node1

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Monday 20 January 2025  06:11:48 -0500 (0:00:00.052)       0:00:17.588 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_pools_global": [],
        "storage_safe_mode_global": true,
        "storage_volumes_global": []
    },
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Monday 20 January 2025  06:11:48 -0500 (0:00:00.034)       0:00:17.623 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:11:48 -0500 (0:00:00.085)       0:00:17.708 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:11:49 -0500 (0:00:00.055)       0:00:17.764 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:11:49 -0500 (0:00:00.048)       0:00:17.812 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:11:49 -0500 (0:00:00.076)       0:00:17.889 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:11:49 -0500 (0:00:00.072)       0:00:17.961 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:11:49 -0500 (0:00:00.076)       0:00:18.037 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:11:49 -0500 (0:00:00.037)       0:00:18.075 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:11:49 -0500 (0:00:00.026)       0:00:18.102 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:11:49 -0500 (0:00:00.078)       0:00:18.180 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:11:52 -0500 (0:00:02.944)       0:00:21.125 ******** 
ok: [managed-node1] => {
    "storage_pools": []
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:11:52 -0500 (0:00:00.047)       0:00:21.172 ******** 
ok: [managed-node1] => {
    "storage_volumes": [
        {
            "disks": [
                "sda"
            ],
            "encryption": true,
            "encryption_luks_version": "luks2",
            "mount_point": "/opt/test1",
            "name": "foo",
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:11:52 -0500 (0:00:00.045)       0:00:21.218 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "cryptsetup"
    ],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:11:56 -0500 (0:00:03.857)       0:00:25.075 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:11:56 -0500 (0:00:00.038)       0:00:25.114 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:11:56 -0500 (0:00:00.020)       0:00:25.134 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:11:56 -0500 (0:00:00.021)       0:00:25.155 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:11:56 -0500 (0:00:00.020)       0:00:25.176 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:11:59 -0500 (0:00:02.950)       0:00:28.126 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:12:01 -0500 (0:00:01.774)       0:00:29.901 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:12:01 -0500 (0:00:00.037)       0:00:29.938 ******** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:12:01 -0500 (0:00:00.018)       0:00:29.956 ******** 
fatal: [managed-node1]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

encrypted volume 'foo' missing key/password

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Monday 20 January 2025  06:12:05 -0500 (0:00:03.939)       0:00:33.896 ******** 
fatal: [managed-node1]: FAILED! => {
    "changed": false
}

MSG:

{'msg': "encrypted volume 'foo' missing key/password", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [], 'volumes': [{'encryption': True, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': 'luks2', 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'foo', 'raid_level': None, 'size': 10737418240, 'state': 'present', 'type': 'disk', 'disks': ['sda'], 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_stripe_size': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'grow_to_fill': False, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, 'shared': False}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:12:05 -0500 (0:00:00.031)       0:00:33.928 ******** 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Monday 20 January 2025  06:12:05 -0500 (0:00:00.018)       0:00:33.946 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Monday 20 January 2025  06:12:05 -0500 (0:00:00.022)       0:00:33.968 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Monday 20 January 2025  06:12:05 -0500 (0:00:00.027)       0:00:33.996 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Create an encrypted disk volume w/ default fs] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:101
Monday 20 January 2025  06:12:05 -0500 (0:00:00.019)       0:00:34.015 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:12:05 -0500 (0:00:00.046)       0:00:34.061 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:12:05 -0500 (0:00:00.032)       0:00:34.094 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:12:05 -0500 (0:00:00.096)       0:00:34.191 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:12:05 -0500 (0:00:00.078)       0:00:34.269 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:12:05 -0500 (0:00:00.030)       0:00:34.299 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:12:05 -0500 (0:00:00.029)       0:00:34.329 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:12:05 -0500 (0:00:00.025)       0:00:34.354 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:12:05 -0500 (0:00:00.023)       0:00:34.378 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:12:05 -0500 (0:00:00.051)       0:00:34.429 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:12:08 -0500 (0:00:02.837)       0:00:37.266 ******** 
ok: [managed-node1] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:12:08 -0500 (0:00:00.025)       0:00:37.292 ******** 
ok: [managed-node1] => {
    "storage_volumes": [
        {
            "disks": [
                "sda"
            ],
            "encryption": true,
            "encryption_luks_version": "luks2",
            "encryption_password": "yabbadabbadoo",
            "mount_point": "/opt/test1",
            "name": "foo",
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:12:08 -0500 (0:00:00.033)       0:00:37.326 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "cryptsetup"
    ],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:12:12 -0500 (0:00:03.826)       0:00:41.153 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:12:12 -0500 (0:00:00.068)       0:00:41.221 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:12:12 -0500 (0:00:00.027)       0:00:41.249 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:12:12 -0500 (0:00:00.028)       0:00:41.278 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:12:12 -0500 (0:00:00.027)       0:00:41.305 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:12:15 -0500 (0:00:02.883)       0:00:44.188 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:12:17 -0500 (0:00:01.631)       0:00:45.820 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:12:17 -0500 (0:00:00.034)       0:00:45.855 ******** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:12:17 -0500 (0:00:00.017)       0:00:45.872 ******** 
changed: [managed-node1] => {
    "actions": [
        {
            "action": "create format",
            "device": "/dev/sda",
            "fs_type": "luks"
        },
        {
            "action": "create device",
            "device": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
            "fs_type": "xfs"
        }
    ],
    "changed": true,
    "crypts": [
        {
            "backing_device": "/dev/sda",
            "name": "luks-90b42248-8017-4670-87d9-7e0b25430bab",
            "password": "-",
            "state": "present"
        }
    ],
    "leaves": [
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1",
        "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "xfs",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
            "state": "mounted"
        }
    ],
    "packages": [
        "cryptsetup",
        "xfsprogs"
    ],
    "pools": [],
    "volumes": [
        {
            "_device": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
            "_kernel_device": "/dev/dm-0",
            "_mount_id": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
            "_raw_device": "/dev/sda",
            "_raw_kernel_device": "/dev/sda",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "sda"
            ],
            "encryption": true,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": "luks2",
            "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
            "fs_create_options": "",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "xfs",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_group": null,
            "mount_mode": null,
            "mount_options": "defaults",
            "mount_passno": 0,
            "mount_point": "/opt/test1",
            "mount_user": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "raid_stripe_size": null,
            "size": 10737418240,
            "state": "present",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Monday 20 January 2025  06:12:31 -0500 (0:00:14.299)       0:01:00.171 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Monday 20 January 2025  06:12:31 -0500 (0:00:00.020)       0:01:00.192 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371482.0043535,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "ab8070345774adad92683e9645714452be7be474",
        "ctime": 1737371481.4993525,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 207618244,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1737371481.4993525,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1343,
        "uid": 0,
        "version": "261612921",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Monday 20 January 2025  06:12:31 -0500 (0:00:00.555)       0:01:00.748 ******** 
ok: [managed-node1] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:12:32 -0500 (0:00:00.500)       0:01:01.248 ******** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Monday 20 January 2025  06:12:32 -0500 (0:00:00.018)       0:01:01.266 ******** 
ok: [managed-node1] => {
    "blivet_output": {
        "actions": [
            {
                "action": "create format",
                "device": "/dev/sda",
                "fs_type": "luks"
            },
            {
                "action": "create device",
                "device": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
                "fs_type": "xfs"
            }
        ],
        "changed": true,
        "crypts": [
            {
                "backing_device": "/dev/sda",
                "name": "luks-90b42248-8017-4670-87d9-7e0b25430bab",
                "password": "-",
                "state": "present"
            }
        ],
        "failed": false,
        "leaves": [
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1",
            "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "xfs",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
                "state": "mounted"
            }
        ],
        "packages": [
            "cryptsetup",
            "xfsprogs"
        ],
        "pools": [],
        "volumes": [
            {
                "_device": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
                "_kernel_device": "/dev/dm-0",
                "_mount_id": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
                "_raw_device": "/dev/sda",
                "_raw_kernel_device": "/dev/sda",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "sda"
                ],
                "encryption": true,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": "luks2",
                "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                "fs_create_options": "",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "xfs",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_group": null,
                "mount_mode": null,
                "mount_options": "defaults",
                "mount_passno": 0,
                "mount_point": "/opt/test1",
                "mount_user": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "raid_stripe_size": null,
                "size": 10737418240,
                "state": "present",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Monday 20 January 2025  06:12:32 -0500 (0:00:00.031)       0:01:01.298 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Monday 20 January 2025  06:12:32 -0500 (0:00:00.034)       0:01:01.332 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": [
            {
                "_device": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
                "_kernel_device": "/dev/dm-0",
                "_mount_id": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
                "_raw_device": "/dev/sda",
                "_raw_kernel_device": "/dev/sda",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "sda"
                ],
                "encryption": true,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": "luks2",
                "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                "fs_create_options": "",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "xfs",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_group": null,
                "mount_mode": null,
                "mount_options": "defaults",
                "mount_passno": 0,
                "mount_point": "/opt/test1",
                "mount_user": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "raid_stripe_size": null,
                "size": 10737418240,
                "state": "present",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Monday 20 January 2025  06:12:32 -0500 (0:00:00.030)       0:01:01.362 ******** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Monday 20 January 2025  06:12:32 -0500 (0:00:00.021)       0:01:01.384 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Monday 20 January 2025  06:12:33 -0500 (0:00:00.870)       0:01:02.254 ******** 
changed: [managed-node1] => (item={'src': '/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Monday 20 January 2025  06:12:33 -0500 (0:00:00.501)       0:01:02.756 ******** 
skipping: [managed-node1] => (item={'src': '/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Monday 20 January 2025  06:12:34 -0500 (0:00:00.028)       0:01:02.784 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Monday 20 January 2025  06:12:34 -0500 (0:00:00.625)       0:01:03.409 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371407.2601843,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1737371405.34018,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 283116674,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1737371405.33918,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1693684562",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Monday 20 January 2025  06:12:35 -0500 (0:00:00.480)       0:01:03.890 ******** 
changed: [managed-node1] => (item={'backing_device': '/dev/sda', 'name': 'luks-90b42248-8017-4670-87d9-7e0b25430bab', 'password': '-', 'state': 'present'}) => {
    "ansible_loop_var": "entry",
    "backup": "",
    "changed": true,
    "entry": {
        "backing_device": "/dev/sda",
        "name": "luks-90b42248-8017-4670-87d9-7e0b25430bab",
        "password": "-",
        "state": "present"
    }
}

MSG:

line added

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Monday 20 January 2025  06:12:35 -0500 (0:00:00.562)       0:01:04.453 ******** 
ok: [managed-node1]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:114
Monday 20 January 2025  06:12:36 -0500 (0:00:00.987)       0:01:05.440 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node1

TASK [Print out pool information] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Monday 20 January 2025  06:12:36 -0500 (0:00:00.098)       0:01:05.539 ******** 
skipping: [managed-node1] => {}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Monday 20 January 2025  06:12:36 -0500 (0:00:00.063)       0:01:05.602 ******** 
ok: [managed-node1] => {
    "_storage_volumes_list": [
        {
            "_device": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
            "_kernel_device": "/dev/dm-0",
            "_mount_id": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
            "_raw_device": "/dev/sda",
            "_raw_kernel_device": "/dev/sda",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "sda"
            ],
            "encryption": true,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": "luks2",
            "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
            "fs_create_options": "",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "xfs",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_group": null,
            "mount_mode": null,
            "mount_options": "defaults",
            "mount_passno": 0,
            "mount_point": "/opt/test1",
            "mount_user": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "raid_stripe_size": null,
            "size": 10737418240,
            "state": "present",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Monday 20 January 2025  06:12:36 -0500 (0:00:00.092)       0:01:05.695 ******** 
ok: [managed-node1] => {
    "changed": false,
    "info": {
        "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
            "size": "10G",
            "type": "crypt",
            "uuid": "4b109b06-6df0-401a-9819-70220ebceab9"
        },
        "/dev/sda": {
            "fstype": "crypto_LUKS",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "90b42248-8017-4670-87d9-7e0b25430bab"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Monday 20 January 2025  06:12:37 -0500 (0:00:00.614)       0:01:06.309 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.002513",
    "end": "2025-01-20 06:12:38.006249",
    "rc": 0,
    "start": "2025-01-20 06:12:38.003736"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Monday 20 January 2025  06:12:38 -0500 (0:00:00.599)       0:01:06.909 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002449",
    "end": "2025-01-20 06:12:38.732446",
    "failed_when_result": false,
    "rc": 0,
    "start": "2025-01-20 06:12:38.729997"
}

STDOUT:

luks-90b42248-8017-4670-87d9-7e0b25430bab /dev/sda -

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Monday 20 January 2025  06:12:38 -0500 (0:00:00.690)       0:01:07.599 ******** 

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Monday 20 January 2025  06:12:38 -0500 (0:00:00.037)       0:01:07.637 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node1

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Monday 20 January 2025  06:12:38 -0500 (0:00:00.117)       0:01:07.754 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Monday 20 January 2025  06:12:39 -0500 (0:00:00.081)       0:01:07.836 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node1

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Monday 20 January 2025  06:12:39 -0500 (0:00:00.248)       0:01:08.084 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Monday 20 January 2025  06:12:39 -0500 (0:00:00.040)       0:01:08.125 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Monday 20 January 2025  06:12:39 -0500 (0:00:00.039)       0:01:08.164 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Monday 20 January 2025  06:12:39 -0500 (0:00:00.031)       0:01:08.196 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Monday 20 January 2025  06:12:39 -0500 (0:00:00.052)       0:01:08.249 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Monday 20 January 2025  06:12:39 -0500 (0:00:00.028)       0:01:08.277 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Monday 20 January 2025  06:12:39 -0500 (0:00:00.037)       0:01:08.315 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Monday 20 January 2025  06:12:39 -0500 (0:00:00.041)       0:01:08.357 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Monday 20 January 2025  06:12:39 -0500 (0:00:00.029)       0:01:08.386 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Monday 20 January 2025  06:12:39 -0500 (0:00:00.029)       0:01:08.415 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Monday 20 January 2025  06:12:39 -0500 (0:00:00.031)       0:01:08.447 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Monday 20 January 2025  06:12:39 -0500 (0:00:00.027)       0:01:08.475 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Monday 20 January 2025  06:12:39 -0500 (0:00:00.060)       0:01:08.535 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Monday 20 January 2025  06:12:39 -0500 (0:00:00.037)       0:01:08.573 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Monday 20 January 2025  06:12:39 -0500 (0:00:00.037)       0:01:08.610 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Monday 20 January 2025  06:12:39 -0500 (0:00:00.028)       0:01:08.638 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Monday 20 January 2025  06:12:39 -0500 (0:00:00.040)       0:01:08.678 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Monday 20 January 2025  06:12:39 -0500 (0:00:00.046)       0:01:08.725 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Monday 20 January 2025  06:12:40 -0500 (0:00:00.048)       0:01:08.774 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Monday 20 January 2025  06:12:40 -0500 (0:00:00.045)       0:01:08.819 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371551.1625102,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1737371551.1625102,
        "dev": 6,
        "device_type": 2048,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 35127,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/blockdevice",
        "mode": "0660",
        "mtime": 1737371551.1625102,
        "nlink": 1,
        "path": "/dev/sda",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Monday 20 January 2025  06:12:40 -0500 (0:00:00.461)       0:01:09.281 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Monday 20 January 2025  06:12:40 -0500 (0:00:00.042)       0:01:09.323 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Monday 20 January 2025  06:12:40 -0500 (0:00:00.039)       0:01:09.362 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Monday 20 January 2025  06:12:40 -0500 (0:00:00.044)       0:01:09.406 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "st_volume_type": "disk"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Monday 20 January 2025  06:12:40 -0500 (0:00:00.032)       0:01:09.439 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Monday 20 January 2025  06:12:40 -0500 (0:00:00.034)       0:01:09.473 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Monday 20 January 2025  06:12:40 -0500 (0:00:00.037)       0:01:09.511 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371551.3015106,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1737371551.3015106,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 586579,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1737371551.3015106,
        "nlink": 1,
        "path": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Monday 20 January 2025  06:12:41 -0500 (0:00:00.454)       0:01:09.966 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Monday 20 January 2025  06:12:44 -0500 (0:00:03.045)       0:01:13.011 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cryptsetup",
        "luksDump",
        "/dev/sda"
    ],
    "delta": "0:00:00.009999",
    "end": "2025-01-20 06:12:44.548631",
    "rc": 0,
    "start": "2025-01-20 06:12:44.538632"
}

STDOUT:

LUKS header information
Version:       	2
Epoch:         	3
Metadata area: 	16384 [bytes]
Keyslots area: 	16744448 [bytes]
UUID:          	90b42248-8017-4670-87d9-7e0b25430bab
Label:         	(no label)
Subsystem:     	(no subsystem)
Flags:       	(no flags)

Data segments:
  0: crypt
	offset: 16777216 [bytes]
	length: (whole device)
	cipher: aes-xts-plain64
	sector: 512 [bytes]

Keyslots:
  0: luks2
	Key:        512 bits
	Priority:   normal
	Cipher:     aes-xts-plain64
	Cipher key: 512 bits
	PBKDF:      argon2i
	Time cost:  4
	Memory:     786224
	Threads:    2
	Salt:       8f d7 96 f3 4b 21 b9 8b aa 1c ab 16 70 f6 7a ec 
	            e2 c6 93 67 4c 7e 90 44 43 59 62 cc ea a7 c3 1a 
	AF stripes: 4000
	AF hash:    sha256
	Area offset:32768 [bytes]
	Area length:258048 [bytes]
	Digest ID:  0
Tokens:
Digests:
  0: pbkdf2
	Hash:       sha256
	Iterations: 120029
	Salt:       8d a8 db 22 8b b1 1a a5 1b 82 19 b7 2d cd ad d3 
	            07 b4 01 d9 45 04 39 87 2c 28 83 2c ee 2c 5d c9 
	Digest:     60 4f cc d4 9f 31 f7 2e 32 65 a2 7c 7b 1c 7c 4b 
	            c5 13 b3 3b 06 b0 ad e0 04 d3 7b 68 7a e6 ac 74 

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Monday 20 January 2025  06:12:44 -0500 (0:00:00.356)       0:01:13.368 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Monday 20 January 2025  06:12:44 -0500 (0:00:00.026)       0:01:13.395 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Monday 20 January 2025  06:12:44 -0500 (0:00:00.028)       0:01:13.423 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Monday 20 January 2025  06:12:44 -0500 (0:00:00.024)       0:01:13.448 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Monday 20 January 2025  06:12:44 -0500 (0:00:00.025)       0:01:13.473 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Monday 20 January 2025  06:12:44 -0500 (0:00:00.070)       0:01:13.544 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Monday 20 January 2025  06:12:44 -0500 (0:00:00.022)       0:01:13.566 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Monday 20 January 2025  06:12:44 -0500 (0:00:00.021)       0:01:13.587 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [
            "luks-90b42248-8017-4670-87d9-7e0b25430bab /dev/sda -"
        ],
        "_storage_test_expected_crypttab_entries": "1",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Monday 20 January 2025  06:12:44 -0500 (0:00:00.026)       0:01:13.613 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Monday 20 January 2025  06:12:44 -0500 (0:00:00.023)       0:01:13.637 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Monday 20 January 2025  06:12:44 -0500 (0:00:00.027)       0:01:13.664 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Monday 20 January 2025  06:12:44 -0500 (0:00:00.026)       0:01:13.690 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Monday 20 January 2025  06:12:44 -0500 (0:00:00.025)       0:01:13.716 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Monday 20 January 2025  06:12:44 -0500 (0:00:00.018)       0:01:13.734 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Monday 20 January 2025  06:12:44 -0500 (0:00:00.018)       0:01:13.752 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Monday 20 January 2025  06:12:45 -0500 (0:00:00.018)       0:01:13.771 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Monday 20 January 2025  06:12:45 -0500 (0:00:00.020)       0:01:13.791 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Monday 20 January 2025  06:12:45 -0500 (0:00:00.020)       0:01:13.811 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Monday 20 January 2025  06:12:45 -0500 (0:00:00.019)       0:01:13.831 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Monday 20 January 2025  06:12:45 -0500 (0:00:00.019)       0:01:13.851 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Monday 20 January 2025  06:12:45 -0500 (0:00:00.019)       0:01:13.871 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Monday 20 January 2025  06:12:45 -0500 (0:00:00.019)       0:01:13.890 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Monday 20 January 2025  06:12:45 -0500 (0:00:00.031)       0:01:13.921 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Monday 20 January 2025  06:12:45 -0500 (0:00:00.022)       0:01:13.943 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Monday 20 January 2025  06:12:45 -0500 (0:00:00.020)       0:01:13.964 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Monday 20 January 2025  06:12:45 -0500 (0:00:00.020)       0:01:13.984 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Monday 20 January 2025  06:12:45 -0500 (0:00:00.025)       0:01:14.010 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Monday 20 January 2025  06:12:45 -0500 (0:00:00.024)       0:01:14.035 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Monday 20 January 2025  06:12:45 -0500 (0:00:00.033)       0:01:14.068 ******** 
skipping: [managed-node1] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Monday 20 January 2025  06:12:45 -0500 (0:00:00.035)       0:01:14.104 ******** 
skipping: [managed-node1] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Monday 20 January 2025  06:12:45 -0500 (0:00:00.035)       0:01:14.140 ******** 
skipping: [managed-node1] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Monday 20 January 2025  06:12:45 -0500 (0:00:00.034)       0:01:14.175 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Monday 20 January 2025  06:12:45 -0500 (0:00:00.030)       0:01:14.206 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Monday 20 January 2025  06:12:45 -0500 (0:00:00.024)       0:01:14.230 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Monday 20 January 2025  06:12:45 -0500 (0:00:00.022)       0:01:14.253 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Monday 20 January 2025  06:12:45 -0500 (0:00:00.021)       0:01:14.274 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Monday 20 January 2025  06:12:45 -0500 (0:00:00.039)       0:01:14.314 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Monday 20 January 2025  06:12:45 -0500 (0:00:00.030)       0:01:14.344 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Monday 20 January 2025  06:12:45 -0500 (0:00:00.032)       0:01:14.377 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Monday 20 January 2025  06:12:45 -0500 (0:00:00.031)       0:01:14.408 ******** 
skipping: [managed-node1] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Monday 20 January 2025  06:12:45 -0500 (0:00:00.035)       0:01:14.444 ******** 
skipping: [managed-node1] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Monday 20 January 2025  06:12:45 -0500 (0:00:00.028)       0:01:14.472 ******** 
skipping: [managed-node1] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Monday 20 January 2025  06:12:45 -0500 (0:00:00.019)       0:01:14.492 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Monday 20 January 2025  06:12:45 -0500 (0:00:00.023)       0:01:14.515 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Monday 20 January 2025  06:12:45 -0500 (0:00:00.022)       0:01:14.538 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Monday 20 January 2025  06:12:45 -0500 (0:00:00.021)       0:01:14.560 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Monday 20 January 2025  06:12:45 -0500 (0:00:00.023)       0:01:14.583 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Monday 20 January 2025  06:12:45 -0500 (0:00:00.019)       0:01:14.603 ******** 
ok: [managed-node1] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Monday 20 January 2025  06:12:45 -0500 (0:00:00.022)       0:01:14.625 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Monday 20 January 2025  06:12:45 -0500 (0:00:00.022)       0:01:14.648 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Monday 20 January 2025  06:12:45 -0500 (0:00:00.021)       0:01:14.670 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Monday 20 January 2025  06:12:45 -0500 (0:00:00.019)       0:01:14.689 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Monday 20 January 2025  06:12:45 -0500 (0:00:00.022)       0:01:14.712 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Monday 20 January 2025  06:12:45 -0500 (0:00:00.020)       0:01:14.733 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Monday 20 January 2025  06:12:45 -0500 (0:00:00.019)       0:01:14.753 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Monday 20 January 2025  06:12:46 -0500 (0:00:00.019)       0:01:14.772 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Monday 20 January 2025  06:12:46 -0500 (0:00:00.019)       0:01:14.792 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Monday 20 January 2025  06:12:46 -0500 (0:00:00.018)       0:01:14.811 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Monday 20 January 2025  06:12:46 -0500 (0:00:00.020)       0:01:14.832 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Create a file] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml:12
Monday 20 January 2025  06:12:46 -0500 (0:00:00.018)       0:01:14.850 ******** 
changed: [managed-node1] => {
    "changed": true,
    "dest": "/opt/test1/quux",
    "gid": 0,
    "group": "root",
    "mode": "0644",
    "owner": "root",
    "secontext": "unconfined_u:object_r:unlabeled_t:s0",
    "size": 0,
    "state": "file",
    "uid": 0
}

TASK [Test for correct handling of safe_mode] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:120
Monday 20 January 2025  06:12:46 -0500 (0:00:00.679)       0:01:15.529 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node1

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Monday 20 January 2025  06:12:46 -0500 (0:00:00.069)       0:01:15.599 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_pools_global": [],
        "storage_safe_mode_global": true,
        "storage_volumes_global": []
    },
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Monday 20 January 2025  06:12:46 -0500 (0:00:00.036)       0:01:15.635 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:12:46 -0500 (0:00:00.079)       0:01:15.715 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:12:46 -0500 (0:00:00.045)       0:01:15.760 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:12:47 -0500 (0:00:00.034)       0:01:15.794 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:12:47 -0500 (0:00:00.075)       0:01:15.870 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:12:47 -0500 (0:00:00.030)       0:01:15.900 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:12:47 -0500 (0:00:00.030)       0:01:15.930 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:12:47 -0500 (0:00:00.027)       0:01:15.958 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:12:47 -0500 (0:00:00.027)       0:01:15.985 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:12:47 -0500 (0:00:00.090)       0:01:16.075 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:12:50 -0500 (0:00:02.887)       0:01:18.963 ******** 
ok: [managed-node1] => {
    "storage_pools": []
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:12:50 -0500 (0:00:00.027)       0:01:18.991 ******** 
ok: [managed-node1] => {
    "storage_volumes": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_luks_version": "luks2",
            "encryption_password": "yabbadabbadoo",
            "mount_point": "/opt/test1",
            "name": "foo",
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:12:50 -0500 (0:00:00.026)       0:01:19.018 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:12:54 -0500 (0:00:03.982)       0:01:23.000 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:12:54 -0500 (0:00:00.037)       0:01:23.038 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:12:54 -0500 (0:00:00.018)       0:01:23.056 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:12:54 -0500 (0:00:00.019)       0:01:23.076 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:12:54 -0500 (0:00:00.018)       0:01:23.095 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:12:57 -0500 (0:00:02.839)       0:01:25.935 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:12:58 -0500 (0:00:01.662)       0:01:27.597 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:12:58 -0500 (0:00:00.052)       0:01:27.650 ******** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:12:58 -0500 (0:00:00.026)       0:01:27.677 ******** 
fatal: [managed-node1]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

cannot remove existing formatting on device 'luks-90b42248-8017-4670-87d9-7e0b25430bab' in safe mode due to encryption removal

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Monday 20 January 2025  06:13:02 -0500 (0:00:03.988)       0:01:31.666 ******** 
fatal: [managed-node1]: FAILED! => {
    "changed": false
}

MSG:

{'msg': "cannot remove existing formatting on device 'luks-90b42248-8017-4670-87d9-7e0b25430bab' in safe mode due to encryption removal", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [], 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': 0, 'encryption_luks_version': 'luks2', 'encryption_password': 'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'foo', 'raid_level': None, 'size': 10720641024, 'state': 'present', 'type': 'disk', 'disks': ['sda'], 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_stripe_size': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'grow_to_fill': False, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, 'shared': False}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:13:02 -0500 (0:00:00.041)       0:01:31.707 ******** 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Monday 20 January 2025  06:13:02 -0500 (0:00:00.019)       0:01:31.726 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Monday 20 January 2025  06:13:02 -0500 (0:00:00.025)       0:01:31.752 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Monday 20 January 2025  06:13:03 -0500 (0:00:00.029)       0:01:31.781 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Stat the file] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:11
Monday 20 January 2025  06:13:03 -0500 (0:00:00.019)       0:01:31.800 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371566.6935453,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1737371566.6935453,
        "dev": 64768,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 131,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0644",
        "mtime": 1737371566.6935453,
        "nlink": 1,
        "path": "/opt/test1/quux",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "796788535",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Assert file presence] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:16
Monday 20 January 2025  06:13:03 -0500 (0:00:00.352)       0:01:32.153 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Remove the encryption layer] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:141
Monday 20 January 2025  06:13:03 -0500 (0:00:00.024)       0:01:32.177 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:13:03 -0500 (0:00:00.065)       0:01:32.242 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:13:03 -0500 (0:00:00.038)       0:01:32.281 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:13:03 -0500 (0:00:00.028)       0:01:32.310 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:13:03 -0500 (0:00:00.062)       0:01:32.372 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:13:03 -0500 (0:00:00.023)       0:01:32.396 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:13:03 -0500 (0:00:00.022)       0:01:32.418 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:13:03 -0500 (0:00:00.022)       0:01:32.441 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:13:03 -0500 (0:00:00.019)       0:01:32.461 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:13:03 -0500 (0:00:00.045)       0:01:32.507 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:13:06 -0500 (0:00:02.832)       0:01:35.339 ******** 
ok: [managed-node1] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:13:06 -0500 (0:00:00.022)       0:01:35.361 ******** 
ok: [managed-node1] => {
    "storage_volumes": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_luks_version": "luks2",
            "encryption_password": "yabbadabbadoo",
            "mount_point": "/opt/test1",
            "name": "foo",
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:13:06 -0500 (0:00:00.025)       0:01:35.387 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:13:10 -0500 (0:00:03.933)       0:01:39.320 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:13:10 -0500 (0:00:00.045)       0:01:39.366 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:13:10 -0500 (0:00:00.055)       0:01:39.421 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:13:10 -0500 (0:00:00.021)       0:01:39.443 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:13:10 -0500 (0:00:00.023)       0:01:39.466 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:13:13 -0500 (0:00:02.831)       0:01:42.298 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:13:15 -0500 (0:00:01.609)       0:01:43.908 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:13:15 -0500 (0:00:00.036)       0:01:43.945 ******** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:13:15 -0500 (0:00:00.025)       0:01:43.970 ******** 
changed: [managed-node1] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
            "fs_type": "xfs"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/sda",
            "fs_type": "luks"
        },
        {
            "action": "create format",
            "device": "/dev/sda",
            "fs_type": "xfs"
        }
    ],
    "changed": true,
    "crypts": [
        {
            "backing_device": "/dev/sda",
            "name": "luks-90b42248-8017-4670-87d9-7e0b25430bab",
            "password": "-",
            "state": "absent"
        }
    ],
    "leaves": [
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1"
    ],
    "mounts": [
        {
            "fstype": "xfs",
            "path": "/opt/test1",
            "src": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
            "state": "absent"
        },
        {
            "dump": 0,
            "fstype": "xfs",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "UUID=5f0315ea-71a4-4497-a6bf-c0d580a2e23e",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [],
    "volumes": [
        {
            "_device": "/dev/sda",
            "_kernel_device": "/dev/sda",
            "_mount_id": "UUID=5f0315ea-71a4-4497-a6bf-c0d580a2e23e",
            "_raw_device": "/dev/sda",
            "_raw_kernel_device": "/dev/sda",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": 0,
            "encryption_luks_version": "luks2",
            "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
            "fs_create_options": "",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "xfs",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_group": null,
            "mount_mode": null,
            "mount_options": "defaults",
            "mount_passno": 0,
            "mount_point": "/opt/test1",
            "mount_user": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "raid_stripe_size": null,
            "size": 10720641024,
            "state": "present",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Monday 20 January 2025  06:13:19 -0500 (0:00:04.182)       0:01:48.153 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Monday 20 January 2025  06:13:19 -0500 (0:00:00.052)       0:01:48.205 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371553.9245164,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "fb4fbb9143f9d5d647293c1ed5bdbafb918b7238",
        "ctime": 1737371553.9215164,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 207618244,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1737371553.9215164,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1425,
        "uid": 0,
        "version": "261612921",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Monday 20 January 2025  06:13:19 -0500 (0:00:00.388)       0:01:48.593 ******** 
ok: [managed-node1] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:13:20 -0500 (0:00:00.364)       0:01:48.958 ******** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Monday 20 January 2025  06:13:20 -0500 (0:00:00.024)       0:01:48.983 ******** 
ok: [managed-node1] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
                "fs_type": "xfs"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/sda",
                "fs_type": "luks"
            },
            {
                "action": "create format",
                "device": "/dev/sda",
                "fs_type": "xfs"
            }
        ],
        "changed": true,
        "crypts": [
            {
                "backing_device": "/dev/sda",
                "name": "luks-90b42248-8017-4670-87d9-7e0b25430bab",
                "password": "-",
                "state": "absent"
            }
        ],
        "failed": false,
        "leaves": [
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1"
        ],
        "mounts": [
            {
                "fstype": "xfs",
                "path": "/opt/test1",
                "src": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
                "state": "absent"
            },
            {
                "dump": 0,
                "fstype": "xfs",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "UUID=5f0315ea-71a4-4497-a6bf-c0d580a2e23e",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [],
        "volumes": [
            {
                "_device": "/dev/sda",
                "_kernel_device": "/dev/sda",
                "_mount_id": "UUID=5f0315ea-71a4-4497-a6bf-c0d580a2e23e",
                "_raw_device": "/dev/sda",
                "_raw_kernel_device": "/dev/sda",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": 0,
                "encryption_luks_version": "luks2",
                "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                "fs_create_options": "",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "xfs",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_group": null,
                "mount_mode": null,
                "mount_options": "defaults",
                "mount_passno": 0,
                "mount_point": "/opt/test1",
                "mount_user": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "raid_stripe_size": null,
                "size": 10720641024,
                "state": "present",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Monday 20 January 2025  06:13:20 -0500 (0:00:00.028)       0:01:49.011 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Monday 20 January 2025  06:13:20 -0500 (0:00:00.026)       0:01:49.038 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": [
            {
                "_device": "/dev/sda",
                "_kernel_device": "/dev/sda",
                "_mount_id": "UUID=5f0315ea-71a4-4497-a6bf-c0d580a2e23e",
                "_raw_device": "/dev/sda",
                "_raw_kernel_device": "/dev/sda",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": 0,
                "encryption_luks_version": "luks2",
                "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                "fs_create_options": "",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "xfs",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_group": null,
                "mount_mode": null,
                "mount_options": "defaults",
                "mount_passno": 0,
                "mount_point": "/opt/test1",
                "mount_user": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "raid_stripe_size": null,
                "size": 10720641024,
                "state": "present",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Monday 20 January 2025  06:13:20 -0500 (0:00:00.024)       0:01:49.063 ******** 
changed: [managed-node1] => (item={'src': '/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "fstype": "xfs",
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/luks-90b42248-8017-4670-87d9-7e0b25430bab"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Monday 20 January 2025  06:13:20 -0500 (0:00:00.379)       0:01:49.442 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Monday 20 January 2025  06:13:21 -0500 (0:00:00.620)       0:01:50.063 ******** 
changed: [managed-node1] => (item={'src': 'UUID=5f0315ea-71a4-4497-a6bf-c0d580a2e23e', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "UUID=5f0315ea-71a4-4497-a6bf-c0d580a2e23e",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "UUID=5f0315ea-71a4-4497-a6bf-c0d580a2e23e"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Monday 20 January 2025  06:13:21 -0500 (0:00:00.498)       0:01:50.562 ******** 
skipping: [managed-node1] => (item={'src': 'UUID=5f0315ea-71a4-4497-a6bf-c0d580a2e23e', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "UUID=5f0315ea-71a4-4497-a6bf-c0d580a2e23e",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Monday 20 January 2025  06:13:21 -0500 (0:00:00.047)       0:01:50.609 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Monday 20 January 2025  06:13:22 -0500 (0:00:00.712)       0:01:51.322 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371558.7315273,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "ea56ef50f55d0e41bb0795cb0e0fd001b862cbc2",
        "ctime": 1737371555.5815203,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 192938182,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0600",
        "mtime": 1737371555.5815203,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 53,
        "uid": 0,
        "version": "3531624939",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Monday 20 January 2025  06:13:22 -0500 (0:00:00.415)       0:01:51.738 ******** 
changed: [managed-node1] => (item={'backing_device': '/dev/sda', 'name': 'luks-90b42248-8017-4670-87d9-7e0b25430bab', 'password': '-', 'state': 'absent'}) => {
    "ansible_loop_var": "entry",
    "backup": "",
    "changed": true,
    "entry": {
        "backing_device": "/dev/sda",
        "name": "luks-90b42248-8017-4670-87d9-7e0b25430bab",
        "password": "-",
        "state": "absent"
    },
    "found": 1
}

MSG:

1 line(s) removed

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Monday 20 January 2025  06:13:23 -0500 (0:00:00.518)       0:01:52.256 ******** 
ok: [managed-node1]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:155
Monday 20 January 2025  06:13:24 -0500 (0:00:00.906)       0:01:53.163 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node1

TASK [Print out pool information] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Monday 20 January 2025  06:13:24 -0500 (0:00:00.072)       0:01:53.236 ******** 
skipping: [managed-node1] => {}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Monday 20 January 2025  06:13:24 -0500 (0:00:00.045)       0:01:53.282 ******** 
ok: [managed-node1] => {
    "_storage_volumes_list": [
        {
            "_device": "/dev/sda",
            "_kernel_device": "/dev/sda",
            "_mount_id": "UUID=5f0315ea-71a4-4497-a6bf-c0d580a2e23e",
            "_raw_device": "/dev/sda",
            "_raw_kernel_device": "/dev/sda",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": 0,
            "encryption_luks_version": "luks2",
            "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
            "fs_create_options": "",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "xfs",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_group": null,
            "mount_mode": null,
            "mount_options": "defaults",
            "mount_passno": 0,
            "mount_point": "/opt/test1",
            "mount_user": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "raid_stripe_size": null,
            "size": 10720641024,
            "state": "present",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Monday 20 January 2025  06:13:24 -0500 (0:00:00.075)       0:01:53.357 ******** 
ok: [managed-node1] => {
    "changed": false,
    "info": {
        "/dev/sda": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "5f0315ea-71a4-4497-a6bf-c0d580a2e23e"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Monday 20 January 2025  06:13:25 -0500 (0:00:00.545)       0:01:53.903 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003116",
    "end": "2025-01-20 06:13:25.601109",
    "rc": 0,
    "start": "2025-01-20 06:13:25.597993"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
UUID=5f0315ea-71a4-4497-a6bf-c0d580a2e23e /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Monday 20 January 2025  06:13:25 -0500 (0:00:00.570)       0:01:54.473 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002276",
    "end": "2025-01-20 06:13:26.144702",
    "failed_when_result": false,
    "rc": 0,
    "start": "2025-01-20 06:13:26.142426"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Monday 20 January 2025  06:13:26 -0500 (0:00:00.501)       0:01:54.975 ******** 

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Monday 20 January 2025  06:13:26 -0500 (0:00:00.028)       0:01:55.003 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node1

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Monday 20 January 2025  06:13:26 -0500 (0:00:00.059)       0:01:55.063 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Monday 20 January 2025  06:13:26 -0500 (0:00:00.043)       0:01:55.106 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node1

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Monday 20 January 2025  06:13:26 -0500 (0:00:00.155)       0:01:55.262 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/sda"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Monday 20 January 2025  06:13:26 -0500 (0:00:00.038)       0:01:55.301 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Monday 20 January 2025  06:13:26 -0500 (0:00:00.063)       0:01:55.364 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Monday 20 January 2025  06:13:26 -0500 (0:00:00.033)       0:01:55.398 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Monday 20 January 2025  06:13:26 -0500 (0:00:00.048)       0:01:55.446 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Monday 20 January 2025  06:13:26 -0500 (0:00:00.046)       0:01:55.493 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Monday 20 January 2025  06:13:26 -0500 (0:00:00.049)       0:01:55.542 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Monday 20 January 2025  06:13:26 -0500 (0:00:00.033)       0:01:55.575 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Monday 20 January 2025  06:13:26 -0500 (0:00:00.035)       0:01:55.611 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Monday 20 January 2025  06:13:26 -0500 (0:00:00.071)       0:01:55.683 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Monday 20 January 2025  06:13:26 -0500 (0:00:00.044)       0:01:55.728 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Monday 20 January 2025  06:13:26 -0500 (0:00:00.030)       0:01:55.758 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "UUID=5f0315ea-71a4-4497-a6bf-c0d580a2e23e "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Monday 20 January 2025  06:13:27 -0500 (0:00:00.061)       0:01:55.819 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Monday 20 January 2025  06:13:27 -0500 (0:00:00.039)       0:01:55.859 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Monday 20 January 2025  06:13:27 -0500 (0:00:00.036)       0:01:55.895 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Monday 20 January 2025  06:13:27 -0500 (0:00:00.031)       0:01:55.927 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Monday 20 January 2025  06:13:27 -0500 (0:00:00.039)       0:01:55.966 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Monday 20 January 2025  06:13:27 -0500 (0:00:00.031)       0:01:55.998 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Monday 20 January 2025  06:13:27 -0500 (0:00:00.042)       0:01:56.041 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Monday 20 January 2025  06:13:27 -0500 (0:00:00.045)       0:01:56.087 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371599.2746148,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1737371599.2746148,
        "dev": 6,
        "device_type": 2048,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 35127,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/blockdevice",
        "mode": "0660",
        "mtime": 1737371599.2746148,
        "nlink": 1,
        "path": "/dev/sda",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Monday 20 January 2025  06:13:27 -0500 (0:00:00.474)       0:01:56.562 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Monday 20 January 2025  06:13:27 -0500 (0:00:00.038)       0:01:56.600 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Monday 20 January 2025  06:13:27 -0500 (0:00:00.032)       0:01:56.633 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Monday 20 January 2025  06:13:27 -0500 (0:00:00.036)       0:01:56.669 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "st_volume_type": "disk"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Monday 20 January 2025  06:13:27 -0500 (0:00:00.035)       0:01:56.705 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Monday 20 January 2025  06:13:27 -0500 (0:00:00.043)       0:01:56.749 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Monday 20 January 2025  06:13:28 -0500 (0:00:00.035)       0:01:56.784 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Monday 20 January 2025  06:13:28 -0500 (0:00:00.034)       0:01:56.818 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Monday 20 January 2025  06:13:30 -0500 (0:00:02.903)       0:01:59.722 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Monday 20 January 2025  06:13:31 -0500 (0:00:00.055)       0:01:59.777 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Monday 20 January 2025  06:13:31 -0500 (0:00:00.021)       0:01:59.799 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Monday 20 January 2025  06:13:31 -0500 (0:00:00.042)       0:01:59.841 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Monday 20 January 2025  06:13:31 -0500 (0:00:00.022)       0:01:59.864 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Monday 20 January 2025  06:13:31 -0500 (0:00:00.022)       0:01:59.886 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Monday 20 January 2025  06:13:31 -0500 (0:00:00.023)       0:01:59.910 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Monday 20 January 2025  06:13:31 -0500 (0:00:00.025)       0:01:59.935 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Monday 20 January 2025  06:13:31 -0500 (0:00:00.020)       0:01:59.956 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Monday 20 January 2025  06:13:31 -0500 (0:00:00.026)       0:01:59.982 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Monday 20 January 2025  06:13:31 -0500 (0:00:00.025)       0:02:00.008 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Monday 20 January 2025  06:13:31 -0500 (0:00:00.020)       0:02:00.028 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Monday 20 January 2025  06:13:31 -0500 (0:00:00.021)       0:02:00.049 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Monday 20 January 2025  06:13:31 -0500 (0:00:00.022)       0:02:00.072 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Monday 20 January 2025  06:13:31 -0500 (0:00:00.019)       0:02:00.091 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Monday 20 January 2025  06:13:31 -0500 (0:00:00.027)       0:02:00.119 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Monday 20 January 2025  06:13:31 -0500 (0:00:00.021)       0:02:00.141 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Monday 20 January 2025  06:13:31 -0500 (0:00:00.035)       0:02:00.176 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Monday 20 January 2025  06:13:31 -0500 (0:00:00.034)       0:02:00.211 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Monday 20 January 2025  06:13:31 -0500 (0:00:00.032)       0:02:00.243 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Monday 20 January 2025  06:13:31 -0500 (0:00:00.032)       0:02:00.275 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Monday 20 January 2025  06:13:31 -0500 (0:00:00.030)       0:02:00.306 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Monday 20 January 2025  06:13:31 -0500 (0:00:00.030)       0:02:00.336 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Monday 20 January 2025  06:13:31 -0500 (0:00:00.030)       0:02:00.366 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Monday 20 January 2025  06:13:31 -0500 (0:00:00.029)       0:02:00.396 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Monday 20 January 2025  06:13:31 -0500 (0:00:00.039)       0:02:00.436 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Monday 20 January 2025  06:13:31 -0500 (0:00:00.044)       0:02:00.481 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Monday 20 January 2025  06:13:31 -0500 (0:00:00.047)       0:02:00.528 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Monday 20 January 2025  06:13:31 -0500 (0:00:00.033)       0:02:00.561 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Monday 20 January 2025  06:13:31 -0500 (0:00:00.033)       0:02:00.595 ******** 
skipping: [managed-node1] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Monday 20 January 2025  06:13:31 -0500 (0:00:00.049)       0:02:00.644 ******** 
skipping: [managed-node1] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Monday 20 January 2025  06:13:31 -0500 (0:00:00.037)       0:02:00.682 ******** 
skipping: [managed-node1] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Monday 20 January 2025  06:13:31 -0500 (0:00:00.043)       0:02:00.726 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Monday 20 January 2025  06:13:32 -0500 (0:00:00.041)       0:02:00.767 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Monday 20 January 2025  06:13:32 -0500 (0:00:00.031)       0:02:00.799 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Monday 20 January 2025  06:13:32 -0500 (0:00:00.028)       0:02:00.827 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Monday 20 January 2025  06:13:32 -0500 (0:00:00.030)       0:02:00.857 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Monday 20 January 2025  06:13:32 -0500 (0:00:00.031)       0:02:00.889 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Monday 20 January 2025  06:13:32 -0500 (0:00:00.024)       0:02:00.914 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Monday 20 January 2025  06:13:32 -0500 (0:00:00.025)       0:02:00.939 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Monday 20 January 2025  06:13:32 -0500 (0:00:00.023)       0:02:00.963 ******** 
skipping: [managed-node1] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Monday 20 January 2025  06:13:32 -0500 (0:00:00.023)       0:02:00.986 ******** 
skipping: [managed-node1] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Monday 20 January 2025  06:13:32 -0500 (0:00:00.020)       0:02:01.007 ******** 
skipping: [managed-node1] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Monday 20 January 2025  06:13:32 -0500 (0:00:00.025)       0:02:01.032 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Monday 20 January 2025  06:13:32 -0500 (0:00:00.020)       0:02:01.052 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Monday 20 January 2025  06:13:32 -0500 (0:00:00.029)       0:02:01.082 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Monday 20 January 2025  06:13:32 -0500 (0:00:00.033)       0:02:01.115 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Monday 20 January 2025  06:13:32 -0500 (0:00:00.034)       0:02:01.150 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Monday 20 January 2025  06:13:32 -0500 (0:00:00.019)       0:02:01.169 ******** 
ok: [managed-node1] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Monday 20 January 2025  06:13:32 -0500 (0:00:00.028)       0:02:01.198 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Monday 20 January 2025  06:13:32 -0500 (0:00:00.028)       0:02:01.226 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Monday 20 January 2025  06:13:32 -0500 (0:00:00.023)       0:02:01.250 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Monday 20 January 2025  06:13:32 -0500 (0:00:00.030)       0:02:01.280 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Monday 20 January 2025  06:13:32 -0500 (0:00:00.025)       0:02:01.305 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Monday 20 January 2025  06:13:32 -0500 (0:00:00.022)       0:02:01.328 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Monday 20 January 2025  06:13:32 -0500 (0:00:00.024)       0:02:01.353 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Monday 20 January 2025  06:13:32 -0500 (0:00:00.021)       0:02:01.374 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Monday 20 January 2025  06:13:32 -0500 (0:00:00.020)       0:02:01.395 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Monday 20 January 2025  06:13:32 -0500 (0:00:00.020)       0:02:01.415 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Monday 20 January 2025  06:13:32 -0500 (0:00:00.019)       0:02:01.435 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Create a file] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml:12
Monday 20 January 2025  06:13:32 -0500 (0:00:00.018)       0:02:01.454 ******** 
changed: [managed-node1] => {
    "changed": true,
    "dest": "/opt/test1/quux",
    "gid": 0,
    "group": "root",
    "mode": "0644",
    "owner": "root",
    "secontext": "unconfined_u:object_r:unlabeled_t:s0",
    "size": 0,
    "state": "file",
    "uid": 0
}

TASK [Test for correct handling of safe_mode] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:161
Monday 20 January 2025  06:13:33 -0500 (0:00:00.366)       0:02:01.820 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node1

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Monday 20 January 2025  06:13:33 -0500 (0:00:00.087)       0:02:01.907 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_pools_global": [],
        "storage_safe_mode_global": true,
        "storage_volumes_global": []
    },
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Monday 20 January 2025  06:13:33 -0500 (0:00:00.037)       0:02:01.945 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:13:33 -0500 (0:00:00.031)       0:02:01.976 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:13:33 -0500 (0:00:00.038)       0:02:02.015 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:13:33 -0500 (0:00:00.032)       0:02:02.047 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:13:33 -0500 (0:00:00.053)       0:02:02.101 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:13:33 -0500 (0:00:00.019)       0:02:02.121 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:13:33 -0500 (0:00:00.021)       0:02:02.142 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:13:33 -0500 (0:00:00.018)       0:02:02.160 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:13:33 -0500 (0:00:00.018)       0:02:02.179 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:13:33 -0500 (0:00:00.050)       0:02:02.229 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:13:36 -0500 (0:00:02.860)       0:02:05.089 ******** 
ok: [managed-node1] => {
    "storage_pools": []
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:13:36 -0500 (0:00:00.039)       0:02:05.128 ******** 
ok: [managed-node1] => {
    "storage_volumes": [
        {
            "disks": [
                "sda"
            ],
            "encryption": true,
            "encryption_luks_version": "luks2",
            "encryption_password": "yabbadabbadoo",
            "mount_point": "/opt/test1",
            "name": "foo",
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:13:36 -0500 (0:00:00.037)       0:02:05.166 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "cryptsetup"
    ],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:13:40 -0500 (0:00:04.095)       0:02:09.261 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:13:40 -0500 (0:00:00.054)       0:02:09.316 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:13:40 -0500 (0:00:00.035)       0:02:09.352 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:13:40 -0500 (0:00:00.031)       0:02:09.383 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:13:40 -0500 (0:00:00.024)       0:02:09.408 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:13:43 -0500 (0:00:02.944)       0:02:12.353 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-cryptsetup@luk...d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service": {
                "name": "systemd-cryptsetup@luk...d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
                "source": "systemd",
                "state": "inactive",
                "status": "generated"
            },
            "systemd-cryptsetup@luks\\x2d90b42248\\x2d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service": {
                "name": "systemd-cryptsetup@luks\\x2d90b42248\\x2d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:13:45 -0500 (0:00:01.742)       0:02:14.095 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": [
            "systemd-cryptsetup@luks\\x2d90b42248\\x2d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
            "systemd-cryptsetup@luk...d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service"
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:13:45 -0500 (0:00:00.113)       0:02:14.208 ******** 
changed: [managed-node1] => (item=systemd-cryptsetup@luks\x2d90b42248\x2d8017\x2d4670\x2d87d9\x2d7e0b25430bab.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luks\\x2d90b42248\\x2d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
    "name": "systemd-cryptsetup@luks\\x2d90b42248\\x2d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "After": "dev-sda.device cryptsetup-pre.target systemd-journald.socket system-systemd\\x2dcryptsetup.slice",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "Before": "umount.target cryptsetup.target",
        "BindsTo": "dev-sda.device",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "yes",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "Conflicts": "umount.target",
        "ControlPID": "0",
        "DefaultDependencies": "no",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "Cryptography Setup for luks-90b42248-8017-4670-87d9-7e0b25430bab",
        "DevicePolicy": "auto",
        "Documentation": "man:crypttab(5) man:systemd-cryptsetup-generator(8) man:systemd-cryptsetup@.service(8)",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-90b42248-8017-4670-87d9-7e0b25430bab /dev/sda -  ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
        "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-90b42248-8017-4670-87d9-7e0b25430bab ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2d90b42248\\x2d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luks\\x2d90b42248\\x2d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
        "IgnoreOnIsolate": "yes",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "shared",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "0",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "65536",
        "LimitMEMLOCKSoft": "65536",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "262144",
        "LimitNOFILESoft": "1024",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadState": "loaded",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luks\\x2d90b42248\\x2d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
        "NeedDaemonReload": "yes",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "yes",
        "RemoveIPC": "no",
        "RequiredBy": "cryptsetup.target",
        "Requires": "system-systemd\\x2dcryptsetup.slice",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "Slice": "system-systemd\\x2dcryptsetup.slice",
        "SourcePath": "/etc/crypttab",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "journal",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestamp": "Mon 2025-01-20 06:13:22 EST",
        "StateChangeTimestampMonotonic": "2420391544",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "infinity",
        "TimeoutStopUSec": "infinity",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "Type": "oneshot",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFilePreset": "disabled",
        "UnitFileState": "generated",
        "UtmpMode": "init",
        "WantedBy": "dev-sda.device",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}
changed: [managed-node1] => (item=systemd-cryptsetup@luk...d8017\x2d4670\x2d87d9\x2d7e0b25430bab.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luk...d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
    "name": "systemd-cryptsetup@luk...d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luk...d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luk...d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.NoSuchUnit \"Unit systemd-cryptsetup@luk...d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service not found.\"",
        "LoadState": "not-found",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luk...d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:13:47 -0500 (0:00:01.612)       0:02:15.821 ******** 
fatal: [managed-node1]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

cannot remove existing formatting on device 'sda' in safe mode due to adding encryption

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Monday 20 January 2025  06:13:51 -0500 (0:00:04.187)       0:02:20.008 ******** 
fatal: [managed-node1]: FAILED! => {
    "changed": false
}

MSG:

{'msg': "cannot remove existing formatting on device 'sda' in safe mode due to adding encryption", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [], 'volumes': [{'encryption': True, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': 'luks2', 'encryption_password': 'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'foo', 'raid_level': None, 'size': 10737418240, 'state': 'present', 'type': 'disk', 'disks': ['sda'], 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_stripe_size': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'grow_to_fill': False, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, 'shared': False}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:13:51 -0500 (0:00:00.037)       0:02:20.046 ******** 
changed: [managed-node1] => (item=systemd-cryptsetup@luks\x2d90b42248\x2d8017\x2d4670\x2d87d9\x2d7e0b25430bab.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luks\\x2d90b42248\\x2d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
    "name": "systemd-cryptsetup@luks\\x2d90b42248\\x2d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luks\\x2d90b42248\\x2d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/dev/null",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luks\\x2d90b42248\\x2d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luks\\x2d90b42248\\x2d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service is masked.\"",
        "LoadState": "masked",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luks\\x2d90b42248\\x2d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFileState": "masked",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}
changed: [managed-node1] => (item=systemd-cryptsetup@luk...d8017\x2d4670\x2d87d9\x2d7e0b25430bab.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luk...d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
    "name": "systemd-cryptsetup@luk...d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luk...d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/dev/null",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luk...d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luk...d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service is masked.\"",
        "LoadState": "masked",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luk...d8017\\x2d4670\\x2d87d9\\x2d7e0b25430bab.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFileState": "masked",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Monday 20 January 2025  06:13:52 -0500 (0:00:01.366)       0:02:21.412 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Monday 20 January 2025  06:13:52 -0500 (0:00:00.035)       0:02:21.448 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Monday 20 January 2025  06:13:52 -0500 (0:00:00.045)       0:02:21.493 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Stat the file] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:11
Monday 20 January 2025  06:13:52 -0500 (0:00:00.029)       0:02:21.522 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371612.9926393,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1737371612.9926393,
        "dev": 2048,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 131,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0644",
        "mtime": 1737371612.9926393,
        "nlink": 1,
        "path": "/opt/test1/quux",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "2387686713",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Assert file presence] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:16
Monday 20 January 2025  06:13:53 -0500 (0:00:00.429)       0:02:21.951 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Add encryption to the volume] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:182
Monday 20 January 2025  06:13:53 -0500 (0:00:00.035)       0:02:21.987 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:13:53 -0500 (0:00:00.100)       0:02:22.088 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:13:53 -0500 (0:00:00.046)       0:02:22.134 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:13:53 -0500 (0:00:00.035)       0:02:22.170 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:13:53 -0500 (0:00:00.085)       0:02:22.255 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:13:53 -0500 (0:00:00.038)       0:02:22.294 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:13:53 -0500 (0:00:00.034)       0:02:22.329 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:13:53 -0500 (0:00:00.031)       0:02:22.360 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:13:53 -0500 (0:00:00.036)       0:02:22.397 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:13:53 -0500 (0:00:00.071)       0:02:22.468 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:13:56 -0500 (0:00:02.939)       0:02:25.407 ******** 
ok: [managed-node1] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:13:56 -0500 (0:00:00.023)       0:02:25.431 ******** 
ok: [managed-node1] => {
    "storage_volumes": [
        {
            "disks": [
                "sda"
            ],
            "encryption": true,
            "encryption_luks_version": "luks2",
            "encryption_password": "yabbadabbadoo",
            "mount_point": "/opt/test1",
            "name": "foo",
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:13:56 -0500 (0:00:00.025)       0:02:25.456 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "cryptsetup"
    ],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:14:00 -0500 (0:00:03.666)       0:02:29.122 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:14:00 -0500 (0:00:00.054)       0:02:29.177 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:14:00 -0500 (0:00:00.030)       0:02:29.208 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:14:00 -0500 (0:00:00.030)       0:02:29.238 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:14:00 -0500 (0:00:00.027)       0:02:29.266 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:14:03 -0500 (0:00:02.906)       0:02:32.172 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:14:05 -0500 (0:00:01.615)       0:02:33.787 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:14:05 -0500 (0:00:00.032)       0:02:33.820 ******** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:14:05 -0500 (0:00:00.019)       0:02:33.840 ******** 
changed: [managed-node1] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/sda",
            "fs_type": "xfs"
        },
        {
            "action": "create format",
            "device": "/dev/sda",
            "fs_type": "luks"
        },
        {
            "action": "create device",
            "device": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
            "fs_type": "xfs"
        }
    ],
    "changed": true,
    "crypts": [
        {
            "backing_device": "/dev/sda",
            "name": "luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
            "password": "-",
            "state": "present"
        }
    ],
    "leaves": [
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1",
        "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f"
    ],
    "mounts": [
        {
            "fstype": "xfs",
            "path": "/opt/test1",
            "src": "UUID=5f0315ea-71a4-4497-a6bf-c0d580a2e23e",
            "state": "absent"
        },
        {
            "dump": 0,
            "fstype": "xfs",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "cryptsetup"
    ],
    "pools": [],
    "volumes": [
        {
            "_device": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
            "_kernel_device": "/dev/dm-0",
            "_mount_id": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
            "_raw_device": "/dev/sda",
            "_raw_kernel_device": "/dev/sda",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "sda"
            ],
            "encryption": true,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": "luks2",
            "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
            "fs_create_options": "",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "xfs",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_group": null,
            "mount_mode": null,
            "mount_options": "defaults",
            "mount_passno": 0,
            "mount_point": "/opt/test1",
            "mount_user": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "raid_stripe_size": null,
            "size": 10737418240,
            "state": "present",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Monday 20 January 2025  06:14:19 -0500 (0:00:14.557)       0:02:48.397 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Monday 20 January 2025  06:14:19 -0500 (0:00:00.033)       0:02:48.431 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371601.7016191,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "c207794dc7a9530cd4bc4d056f62fa9e70a7ccb2",
        "ctime": 1737371601.6986191,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 207618244,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1737371601.6986191,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1413,
        "uid": 0,
        "version": "261612921",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Monday 20 January 2025  06:14:20 -0500 (0:00:00.426)       0:02:48.857 ******** 
ok: [managed-node1] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:14:20 -0500 (0:00:00.422)       0:02:49.280 ******** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Monday 20 January 2025  06:14:20 -0500 (0:00:00.040)       0:02:49.321 ******** 
ok: [managed-node1] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/sda",
                "fs_type": "xfs"
            },
            {
                "action": "create format",
                "device": "/dev/sda",
                "fs_type": "luks"
            },
            {
                "action": "create device",
                "device": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
                "fs_type": "xfs"
            }
        ],
        "changed": true,
        "crypts": [
            {
                "backing_device": "/dev/sda",
                "name": "luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
                "password": "-",
                "state": "present"
            }
        ],
        "failed": false,
        "leaves": [
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1",
            "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f"
        ],
        "mounts": [
            {
                "fstype": "xfs",
                "path": "/opt/test1",
                "src": "UUID=5f0315ea-71a4-4497-a6bf-c0d580a2e23e",
                "state": "absent"
            },
            {
                "dump": 0,
                "fstype": "xfs",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "cryptsetup"
        ],
        "pools": [],
        "volumes": [
            {
                "_device": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
                "_kernel_device": "/dev/dm-0",
                "_mount_id": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
                "_raw_device": "/dev/sda",
                "_raw_kernel_device": "/dev/sda",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "sda"
                ],
                "encryption": true,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": "luks2",
                "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                "fs_create_options": "",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "xfs",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_group": null,
                "mount_mode": null,
                "mount_options": "defaults",
                "mount_passno": 0,
                "mount_point": "/opt/test1",
                "mount_user": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "raid_stripe_size": null,
                "size": 10737418240,
                "state": "present",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Monday 20 January 2025  06:14:20 -0500 (0:00:00.040)       0:02:49.362 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Monday 20 January 2025  06:14:20 -0500 (0:00:00.086)       0:02:49.448 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": [
            {
                "_device": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
                "_kernel_device": "/dev/dm-0",
                "_mount_id": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
                "_raw_device": "/dev/sda",
                "_raw_kernel_device": "/dev/sda",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "sda"
                ],
                "encryption": true,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": "luks2",
                "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                "fs_create_options": "",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "xfs",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_group": null,
                "mount_mode": null,
                "mount_options": "defaults",
                "mount_passno": 0,
                "mount_point": "/opt/test1",
                "mount_user": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "raid_stripe_size": null,
                "size": 10737418240,
                "state": "present",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Monday 20 January 2025  06:14:20 -0500 (0:00:00.036)       0:02:49.484 ******** 
changed: [managed-node1] => (item={'src': 'UUID=5f0315ea-71a4-4497-a6bf-c0d580a2e23e', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "fstype": "xfs",
        "path": "/opt/test1",
        "src": "UUID=5f0315ea-71a4-4497-a6bf-c0d580a2e23e",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "UUID=5f0315ea-71a4-4497-a6bf-c0d580a2e23e"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Monday 20 January 2025  06:14:21 -0500 (0:00:00.604)       0:02:50.089 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Monday 20 January 2025  06:14:22 -0500 (0:00:00.751)       0:02:50.840 ******** 
changed: [managed-node1] => (item={'src': '/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Monday 20 January 2025  06:14:22 -0500 (0:00:00.487)       0:02:51.328 ******** 
skipping: [managed-node1] => (item={'src': '/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Monday 20 January 2025  06:14:22 -0500 (0:00:00.042)       0:02:51.370 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Monday 20 January 2025  06:14:23 -0500 (0:00:00.663)       0:02:52.034 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371606.1436272,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1737371603.333622,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 312475844,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1737371603.3316221,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "539867588",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Monday 20 January 2025  06:14:23 -0500 (0:00:00.397)       0:02:52.431 ******** 
changed: [managed-node1] => (item={'backing_device': '/dev/sda', 'name': 'luks-45ea6580-705d-46a9-b81a-78f69257cc6f', 'password': '-', 'state': 'present'}) => {
    "ansible_loop_var": "entry",
    "backup": "",
    "changed": true,
    "entry": {
        "backing_device": "/dev/sda",
        "name": "luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
        "password": "-",
        "state": "present"
    }
}

MSG:

line added

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Monday 20 January 2025  06:14:24 -0500 (0:00:00.365)       0:02:52.797 ******** 
ok: [managed-node1]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:196
Monday 20 January 2025  06:14:24 -0500 (0:00:00.750)       0:02:53.548 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node1

TASK [Print out pool information] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Monday 20 January 2025  06:14:24 -0500 (0:00:00.047)       0:02:53.596 ******** 
skipping: [managed-node1] => {}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Monday 20 January 2025  06:14:24 -0500 (0:00:00.021)       0:02:53.617 ******** 
ok: [managed-node1] => {
    "_storage_volumes_list": [
        {
            "_device": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
            "_kernel_device": "/dev/dm-0",
            "_mount_id": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
            "_raw_device": "/dev/sda",
            "_raw_kernel_device": "/dev/sda",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "sda"
            ],
            "encryption": true,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": "luks2",
            "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
            "fs_create_options": "",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "xfs",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_group": null,
            "mount_mode": null,
            "mount_options": "defaults",
            "mount_passno": 0,
            "mount_point": "/opt/test1",
            "mount_user": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "raid_stripe_size": null,
            "size": 10737418240,
            "state": "present",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Monday 20 January 2025  06:14:24 -0500 (0:00:00.026)       0:02:53.644 ******** 
ok: [managed-node1] => {
    "changed": false,
    "info": {
        "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
            "size": "10G",
            "type": "crypt",
            "uuid": "46d2016a-2dcd-4263-b2b9-95a68d74139b"
        },
        "/dev/sda": {
            "fstype": "crypto_LUKS",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "45ea6580-705d-46a9-b81a-78f69257cc6f"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Monday 20 January 2025  06:14:25 -0500 (0:00:00.356)       0:02:54.000 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.002546",
    "end": "2025-01-20 06:14:25.532186",
    "rc": 0,
    "start": "2025-01-20 06:14:25.529640"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Monday 20 January 2025  06:14:25 -0500 (0:00:00.356)       0:02:54.357 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002450",
    "end": "2025-01-20 06:14:25.892237",
    "failed_when_result": false,
    "rc": 0,
    "start": "2025-01-20 06:14:25.889787"
}

STDOUT:

luks-45ea6580-705d-46a9-b81a-78f69257cc6f /dev/sda -

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Monday 20 January 2025  06:14:25 -0500 (0:00:00.366)       0:02:54.724 ******** 

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Monday 20 January 2025  06:14:25 -0500 (0:00:00.028)       0:02:54.753 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node1

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Monday 20 January 2025  06:14:26 -0500 (0:00:00.057)       0:02:54.810 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Monday 20 January 2025  06:14:26 -0500 (0:00:00.037)       0:02:54.848 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node1

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Monday 20 January 2025  06:14:26 -0500 (0:00:00.150)       0:02:54.998 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Monday 20 January 2025  06:14:26 -0500 (0:00:00.036)       0:02:55.035 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Monday 20 January 2025  06:14:26 -0500 (0:00:00.079)       0:02:55.115 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Monday 20 January 2025  06:14:26 -0500 (0:00:00.035)       0:02:55.151 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Monday 20 January 2025  06:14:26 -0500 (0:00:00.038)       0:02:55.189 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Monday 20 January 2025  06:14:26 -0500 (0:00:00.033)       0:02:55.223 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Monday 20 January 2025  06:14:26 -0500 (0:00:00.030)       0:02:55.253 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Monday 20 January 2025  06:14:26 -0500 (0:00:00.033)       0:02:55.287 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Monday 20 January 2025  06:14:26 -0500 (0:00:00.035)       0:02:55.322 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Monday 20 January 2025  06:14:26 -0500 (0:00:00.032)       0:02:55.355 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Monday 20 January 2025  06:14:26 -0500 (0:00:00.033)       0:02:55.388 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Monday 20 January 2025  06:14:26 -0500 (0:00:00.032)       0:02:55.421 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Monday 20 January 2025  06:14:26 -0500 (0:00:00.069)       0:02:55.490 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Monday 20 January 2025  06:14:26 -0500 (0:00:00.040)       0:02:55.531 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Monday 20 January 2025  06:14:26 -0500 (0:00:00.067)       0:02:55.599 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Monday 20 January 2025  06:14:26 -0500 (0:00:00.050)       0:02:55.650 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Monday 20 January 2025  06:14:26 -0500 (0:00:00.050)       0:02:55.701 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Monday 20 January 2025  06:14:26 -0500 (0:00:00.031)       0:02:55.733 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Monday 20 January 2025  06:14:27 -0500 (0:00:00.053)       0:02:55.786 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Monday 20 January 2025  06:14:27 -0500 (0:00:00.064)       0:02:55.851 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371659.3727224,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1737371659.3727224,
        "dev": 6,
        "device_type": 2048,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 35127,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/blockdevice",
        "mode": "0660",
        "mtime": 1737371659.3727224,
        "nlink": 1,
        "path": "/dev/sda",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Monday 20 January 2025  06:14:27 -0500 (0:00:00.419)       0:02:56.270 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Monday 20 January 2025  06:14:27 -0500 (0:00:00.039)       0:02:56.310 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Monday 20 January 2025  06:14:27 -0500 (0:00:00.036)       0:02:56.346 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Monday 20 January 2025  06:14:27 -0500 (0:00:00.038)       0:02:56.385 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "st_volume_type": "disk"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Monday 20 January 2025  06:14:27 -0500 (0:00:00.034)       0:02:56.420 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Monday 20 January 2025  06:14:27 -0500 (0:00:00.034)       0:02:56.454 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Monday 20 January 2025  06:14:27 -0500 (0:00:00.038)       0:02:56.493 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371659.5037227,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1737371659.5037227,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 613559,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1737371659.5037227,
        "nlink": 1,
        "path": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Monday 20 January 2025  06:14:28 -0500 (0:00:00.412)       0:02:56.906 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Monday 20 January 2025  06:14:31 -0500 (0:00:02.884)       0:02:59.791 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cryptsetup",
        "luksDump",
        "/dev/sda"
    ],
    "delta": "0:00:00.010439",
    "end": "2025-01-20 06:14:31.364847",
    "rc": 0,
    "start": "2025-01-20 06:14:31.354408"
}

STDOUT:

LUKS header information
Version:       	2
Epoch:         	3
Metadata area: 	16384 [bytes]
Keyslots area: 	16744448 [bytes]
UUID:          	45ea6580-705d-46a9-b81a-78f69257cc6f
Label:         	(no label)
Subsystem:     	(no subsystem)
Flags:       	(no flags)

Data segments:
  0: crypt
	offset: 16777216 [bytes]
	length: (whole device)
	cipher: aes-xts-plain64
	sector: 512 [bytes]

Keyslots:
  0: luks2
	Key:        512 bits
	Priority:   normal
	Cipher:     aes-xts-plain64
	Cipher key: 512 bits
	PBKDF:      argon2i
	Time cost:  4
	Memory:     796257
	Threads:    2
	Salt:       ba 93 2d cc b5 ab 4b 0b 9f 83 89 69 a4 d5 6d 95 
	            5c cb 2b 75 27 a5 ce 68 83 f8 c3 9a 7a 6c fb 4f 
	AF stripes: 4000
	AF hash:    sha256
	Area offset:32768 [bytes]
	Area length:258048 [bytes]
	Digest ID:  0
Tokens:
Digests:
  0: pbkdf2
	Hash:       sha256
	Iterations: 109775
	Salt:       af c5 26 88 57 56 26 95 de 00 d3 65 10 90 68 cf 
	            9a 30 f6 56 21 c6 70 fb 05 7e 21 64 46 c9 7a d0 
	Digest:     89 84 bc 9e 92 b1 ea 40 b2 25 e2 44 d7 e4 42 53 
	            94 3a dc 38 12 dc 7c 59 0c 33 7f 2e 18 db f9 5d 

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Monday 20 January 2025  06:14:31 -0500 (0:00:00.403)       0:03:00.194 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Monday 20 January 2025  06:14:31 -0500 (0:00:00.049)       0:03:00.243 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Monday 20 January 2025  06:14:31 -0500 (0:00:00.051)       0:03:00.295 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Monday 20 January 2025  06:14:31 -0500 (0:00:00.037)       0:03:00.333 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Monday 20 January 2025  06:14:31 -0500 (0:00:00.044)       0:03:00.377 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Monday 20 January 2025  06:14:31 -0500 (0:00:00.048)       0:03:00.426 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Monday 20 January 2025  06:14:31 -0500 (0:00:00.033)       0:03:00.459 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Monday 20 January 2025  06:14:31 -0500 (0:00:00.031)       0:03:00.491 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [
            "luks-45ea6580-705d-46a9-b81a-78f69257cc6f /dev/sda -"
        ],
        "_storage_test_expected_crypttab_entries": "1",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Monday 20 January 2025  06:14:31 -0500 (0:00:00.038)       0:03:00.529 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Monday 20 January 2025  06:14:31 -0500 (0:00:00.044)       0:03:00.573 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Monday 20 January 2025  06:14:31 -0500 (0:00:00.037)       0:03:00.611 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Monday 20 January 2025  06:14:31 -0500 (0:00:00.054)       0:03:00.665 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Monday 20 January 2025  06:14:31 -0500 (0:00:00.049)       0:03:00.714 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Monday 20 January 2025  06:14:31 -0500 (0:00:00.037)       0:03:00.752 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Monday 20 January 2025  06:14:32 -0500 (0:00:00.059)       0:03:00.811 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Monday 20 January 2025  06:14:32 -0500 (0:00:00.085)       0:03:00.897 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Monday 20 January 2025  06:14:32 -0500 (0:00:00.033)       0:03:00.930 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Monday 20 January 2025  06:14:32 -0500 (0:00:00.062)       0:03:00.993 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Monday 20 January 2025  06:14:32 -0500 (0:00:00.055)       0:03:01.049 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Monday 20 January 2025  06:14:32 -0500 (0:00:00.032)       0:03:01.081 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Monday 20 January 2025  06:14:32 -0500 (0:00:00.045)       0:03:01.127 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Monday 20 January 2025  06:14:32 -0500 (0:00:00.032)       0:03:01.159 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Monday 20 January 2025  06:14:32 -0500 (0:00:00.030)       0:03:01.190 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Monday 20 January 2025  06:14:32 -0500 (0:00:00.044)       0:03:01.234 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Monday 20 January 2025  06:14:32 -0500 (0:00:00.037)       0:03:01.272 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Monday 20 January 2025  06:14:32 -0500 (0:00:00.035)       0:03:01.307 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Monday 20 January 2025  06:14:32 -0500 (0:00:00.031)       0:03:01.339 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Monday 20 January 2025  06:14:32 -0500 (0:00:00.034)       0:03:01.373 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Monday 20 January 2025  06:14:32 -0500 (0:00:00.033)       0:03:01.407 ******** 
skipping: [managed-node1] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Monday 20 January 2025  06:14:32 -0500 (0:00:00.036)       0:03:01.443 ******** 
skipping: [managed-node1] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Monday 20 January 2025  06:14:32 -0500 (0:00:00.032)       0:03:01.476 ******** 
skipping: [managed-node1] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Monday 20 January 2025  06:14:32 -0500 (0:00:00.038)       0:03:01.514 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Monday 20 January 2025  06:14:32 -0500 (0:00:00.057)       0:03:01.571 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Monday 20 January 2025  06:14:32 -0500 (0:00:00.033)       0:03:01.605 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Monday 20 January 2025  06:14:32 -0500 (0:00:00.039)       0:03:01.644 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Monday 20 January 2025  06:14:32 -0500 (0:00:00.033)       0:03:01.678 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Monday 20 January 2025  06:14:32 -0500 (0:00:00.030)       0:03:01.708 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Monday 20 January 2025  06:14:32 -0500 (0:00:00.047)       0:03:01.756 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Monday 20 January 2025  06:14:33 -0500 (0:00:00.062)       0:03:01.819 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Monday 20 January 2025  06:14:33 -0500 (0:00:00.030)       0:03:01.850 ******** 
skipping: [managed-node1] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Monday 20 January 2025  06:14:33 -0500 (0:00:00.038)       0:03:01.889 ******** 
skipping: [managed-node1] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Monday 20 January 2025  06:14:33 -0500 (0:00:00.074)       0:03:01.963 ******** 
skipping: [managed-node1] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Monday 20 January 2025  06:14:33 -0500 (0:00:00.054)       0:03:02.017 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Monday 20 January 2025  06:14:33 -0500 (0:00:00.059)       0:03:02.077 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Monday 20 January 2025  06:14:33 -0500 (0:00:00.033)       0:03:02.110 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Monday 20 January 2025  06:14:33 -0500 (0:00:00.077)       0:03:02.188 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Monday 20 January 2025  06:14:33 -0500 (0:00:00.085)       0:03:02.274 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Monday 20 January 2025  06:14:33 -0500 (0:00:00.064)       0:03:02.338 ******** 
ok: [managed-node1] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Monday 20 January 2025  06:14:33 -0500 (0:00:00.035)       0:03:02.374 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Monday 20 January 2025  06:14:33 -0500 (0:00:00.034)       0:03:02.409 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Monday 20 January 2025  06:14:33 -0500 (0:00:00.036)       0:03:02.445 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Monday 20 January 2025  06:14:33 -0500 (0:00:00.042)       0:03:02.488 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Monday 20 January 2025  06:14:33 -0500 (0:00:00.029)       0:03:02.517 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Monday 20 January 2025  06:14:33 -0500 (0:00:00.030)       0:03:02.548 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Monday 20 January 2025  06:14:33 -0500 (0:00:00.029)       0:03:02.577 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Monday 20 January 2025  06:14:33 -0500 (0:00:00.031)       0:03:02.609 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Monday 20 January 2025  06:14:33 -0500 (0:00:00.023)       0:03:02.632 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Monday 20 January 2025  06:14:33 -0500 (0:00:00.025)       0:03:02.658 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Monday 20 January 2025  06:14:33 -0500 (0:00:00.023)       0:03:02.681 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Test for correct handling of new encrypted volume w/ no key] *************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:203
Monday 20 January 2025  06:14:33 -0500 (0:00:00.020)       0:03:02.702 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node1

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Monday 20 January 2025  06:14:33 -0500 (0:00:00.049)       0:03:02.752 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_pools_global": [],
        "storage_safe_mode_global": true,
        "storage_volumes_global": []
    },
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Monday 20 January 2025  06:14:34 -0500 (0:00:00.029)       0:03:02.782 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:14:34 -0500 (0:00:00.030)       0:03:02.813 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:14:34 -0500 (0:00:00.030)       0:03:02.843 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:14:34 -0500 (0:00:00.024)       0:03:02.867 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:14:34 -0500 (0:00:00.063)       0:03:02.931 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:14:34 -0500 (0:00:00.027)       0:03:02.958 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:14:34 -0500 (0:00:00.021)       0:03:02.979 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:14:34 -0500 (0:00:00.020)       0:03:03.000 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:14:34 -0500 (0:00:00.024)       0:03:03.024 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:14:34 -0500 (0:00:00.062)       0:03:03.087 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:14:37 -0500 (0:00:02.897)       0:03:05.984 ******** 
ok: [managed-node1] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "partition",
            "volumes": [
                {
                    "encryption": true,
                    "encryption_luks_version": "luks2",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g",
                    "type": "partition"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:14:37 -0500 (0:00:00.043)       0:03:06.027 ******** 
ok: [managed-node1] => {
    "storage_volumes": []
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:14:37 -0500 (0:00:00.044)       0:03:06.072 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "cryptsetup"
    ],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:14:41 -0500 (0:00:03.941)       0:03:10.014 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:14:41 -0500 (0:00:00.101)       0:03:10.116 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:14:41 -0500 (0:00:00.027)       0:03:10.144 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:14:41 -0500 (0:00:00.042)       0:03:10.186 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:14:41 -0500 (0:00:00.029)       0:03:10.216 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:14:44 -0500 (0:00:02.900)       0:03:13.117 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:14:46 -0500 (0:00:01.663)       0:03:14.781 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:14:46 -0500 (0:00:00.043)       0:03:14.824 ******** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:14:46 -0500 (0:00:00.028)       0:03:14.853 ******** 
fatal: [managed-node1]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

encrypted volume 'test1' missing key/password

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Monday 20 January 2025  06:14:50 -0500 (0:00:04.093)       0:03:18.947 ******** 
fatal: [managed-node1]: FAILED! => {
    "changed": false
}

MSG:

{'msg': "encrypted volume 'test1' missing key/password", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'encryption_clevis_pin': None, 'encryption_tang_url': None, 'encryption_tang_thumbprint': None, 'grow_to_fill': False, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'shared': False, 'state': 'present', 'type': 'partition', 'volumes': [{'encryption': True, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': 'luks2', 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}]}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'grow_to_fill': False, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, 'shared': False}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': False, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:14:50 -0500 (0:00:00.048)       0:03:18.996 ******** 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Monday 20 January 2025  06:14:50 -0500 (0:00:00.047)       0:03:19.043 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Monday 20 January 2025  06:14:50 -0500 (0:00:00.050)       0:03:19.094 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Monday 20 January 2025  06:14:50 -0500 (0:00:00.049)       0:03:19.143 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Create an encrypted partition volume w/ default fs] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:223
Monday 20 January 2025  06:14:50 -0500 (0:00:00.038)       0:03:19.182 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:14:50 -0500 (0:00:00.126)       0:03:19.308 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:14:50 -0500 (0:00:00.036)       0:03:19.345 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:14:50 -0500 (0:00:00.028)       0:03:19.373 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:14:50 -0500 (0:00:00.071)       0:03:19.445 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:14:50 -0500 (0:00:00.031)       0:03:19.476 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:14:50 -0500 (0:00:00.023)       0:03:19.499 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:14:50 -0500 (0:00:00.021)       0:03:19.521 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:14:50 -0500 (0:00:00.020)       0:03:19.541 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:14:50 -0500 (0:00:00.046)       0:03:19.588 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:14:53 -0500 (0:00:02.952)       0:03:22.541 ******** 
ok: [managed-node1] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "partition",
            "volumes": [
                {
                    "encryption": true,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "yabbadabbadoo",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g",
                    "type": "partition"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:14:53 -0500 (0:00:00.081)       0:03:22.622 ******** 
ok: [managed-node1] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:14:53 -0500 (0:00:00.067)       0:03:22.690 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "cryptsetup"
    ],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:14:57 -0500 (0:00:04.001)       0:03:26.691 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:14:57 -0500 (0:00:00.037)       0:03:26.729 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:14:57 -0500 (0:00:00.025)       0:03:26.754 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:14:58 -0500 (0:00:00.026)       0:03:26.781 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:14:58 -0500 (0:00:00.032)       0:03:26.814 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:15:00 -0500 (0:00:02.853)       0:03:29.667 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:15:02 -0500 (0:00:01.777)       0:03:31.444 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:15:02 -0500 (0:00:00.049)       0:03:31.494 ******** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:15:02 -0500 (0:00:00.031)       0:03:31.526 ******** 
changed: [managed-node1] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
            "fs_type": "xfs"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/sda",
            "fs_type": "luks"
        },
        {
            "action": "create format",
            "device": "/dev/sda",
            "fs_type": "disklabel"
        },
        {
            "action": "create device",
            "device": "/dev/sda1",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/sda1",
            "fs_type": "luks"
        },
        {
            "action": "create device",
            "device": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
            "fs_type": "xfs"
        }
    ],
    "changed": true,
    "crypts": [
        {
            "backing_device": "/dev/sda",
            "name": "luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
            "password": "-",
            "state": "absent"
        },
        {
            "backing_device": "/dev/sda1",
            "name": "luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
            "password": "-",
            "state": "present"
        }
    ],
    "leaves": [
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1",
        "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518"
    ],
    "mounts": [
        {
            "fstype": "xfs",
            "path": "/opt/test1",
            "src": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
            "state": "absent"
        },
        {
            "dump": 0,
            "fstype": "xfs",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "cryptsetup"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "partition",
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
                    "_raw_device": "/dev/sda1",
                    "_raw_kernel_device": "/dev/sda1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": true,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "xfs",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "partition",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Monday 20 January 2025  06:15:17 -0500 (0:00:14.819)       0:03:46.345 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Monday 20 January 2025  06:15:17 -0500 (0:00:00.085)       0:03:46.431 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371662.4227278,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "7c1ced10803f3514c96a886912a7d80f8c280d73",
        "ctime": 1737371662.419728,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 207618244,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1737371662.419728,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1425,
        "uid": 0,
        "version": "261612921",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Monday 20 January 2025  06:15:18 -0500 (0:00:00.520)       0:03:46.952 ******** 
ok: [managed-node1] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:15:18 -0500 (0:00:00.595)       0:03:47.547 ******** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Monday 20 January 2025  06:15:18 -0500 (0:00:00.030)       0:03:47.578 ******** 
ok: [managed-node1] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
                "fs_type": "xfs"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/sda",
                "fs_type": "luks"
            },
            {
                "action": "create format",
                "device": "/dev/sda",
                "fs_type": "disklabel"
            },
            {
                "action": "create device",
                "device": "/dev/sda1",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/sda1",
                "fs_type": "luks"
            },
            {
                "action": "create device",
                "device": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
                "fs_type": "xfs"
            }
        ],
        "changed": true,
        "crypts": [
            {
                "backing_device": "/dev/sda",
                "name": "luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
                "password": "-",
                "state": "absent"
            },
            {
                "backing_device": "/dev/sda1",
                "name": "luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
                "password": "-",
                "state": "present"
            }
        ],
        "failed": false,
        "leaves": [
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1",
            "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518"
        ],
        "mounts": [
            {
                "fstype": "xfs",
                "path": "/opt/test1",
                "src": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
                "state": "absent"
            },
            {
                "dump": 0,
                "fstype": "xfs",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "cryptsetup"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "partition",
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
                        "_raw_device": "/dev/sda1",
                        "_raw_kernel_device": "/dev/sda1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": true,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": "luks2",
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "xfs",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "partition",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Monday 20 January 2025  06:15:18 -0500 (0:00:00.042)       0:03:47.620 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "partition",
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
                        "_raw_device": "/dev/sda1",
                        "_raw_kernel_device": "/dev/sda1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": true,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": "luks2",
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "xfs",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "partition",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Monday 20 January 2025  06:15:18 -0500 (0:00:00.045)       0:03:47.666 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Monday 20 January 2025  06:15:18 -0500 (0:00:00.034)       0:03:47.701 ******** 
changed: [managed-node1] => (item={'src': '/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "fstype": "xfs",
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/luks-45ea6580-705d-46a9-b81a-78f69257cc6f"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Monday 20 January 2025  06:15:19 -0500 (0:00:00.417)       0:03:48.119 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Monday 20 January 2025  06:15:20 -0500 (0:00:00.762)       0:03:48.882 ******** 
changed: [managed-node1] => (item={'src': '/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Monday 20 January 2025  06:15:20 -0500 (0:00:00.475)       0:03:49.357 ******** 
skipping: [managed-node1] => (item={'src': '/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Monday 20 January 2025  06:15:20 -0500 (0:00:00.053)       0:03:49.411 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Monday 20 January 2025  06:15:21 -0500 (0:00:00.799)       0:03:50.211 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371665.8917341,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "f3c9b7daa3ad92e0870174d2b93cbf70fb7c32c4",
        "ctime": 1737371663.9717307,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 440402053,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0600",
        "mtime": 1737371663.9707308,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 53,
        "uid": 0,
        "version": "2947634537",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Monday 20 January 2025  06:15:21 -0500 (0:00:00.443)       0:03:50.655 ******** 
changed: [managed-node1] => (item={'backing_device': '/dev/sda', 'name': 'luks-45ea6580-705d-46a9-b81a-78f69257cc6f', 'password': '-', 'state': 'absent'}) => {
    "ansible_loop_var": "entry",
    "backup": "",
    "changed": true,
    "entry": {
        "backing_device": "/dev/sda",
        "name": "luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
        "password": "-",
        "state": "absent"
    },
    "found": 1
}

MSG:

1 line(s) removed
changed: [managed-node1] => (item={'backing_device': '/dev/sda1', 'name': 'luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518', 'password': '-', 'state': 'present'}) => {
    "ansible_loop_var": "entry",
    "backup": "",
    "changed": true,
    "entry": {
        "backing_device": "/dev/sda1",
        "name": "luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
        "password": "-",
        "state": "present"
    }
}

MSG:

line added

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Monday 20 January 2025  06:15:22 -0500 (0:00:00.795)       0:03:51.450 ******** 
ok: [managed-node1]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:241
Monday 20 January 2025  06:15:23 -0500 (0:00:00.817)       0:03:52.267 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node1

TASK [Print out pool information] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Monday 20 January 2025  06:15:23 -0500 (0:00:00.083)       0:03:52.351 ******** 
ok: [managed-node1] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "partition",
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
                    "_raw_device": "/dev/sda1",
                    "_raw_kernel_device": "/dev/sda1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": true,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "xfs",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "partition",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Monday 20 January 2025  06:15:23 -0500 (0:00:00.041)       0:03:52.393 ******** 
skipping: [managed-node1] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Monday 20 January 2025  06:15:23 -0500 (0:00:00.031)       0:03:52.424 ******** 
ok: [managed-node1] => {
    "changed": false,
    "info": {
        "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
            "size": "10G",
            "type": "crypt",
            "uuid": "f722c14f-aeca-411d-a7d0-42bf527e5e58"
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda1": {
            "fstype": "crypto_LUKS",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda1",
            "size": "10G",
            "type": "partition",
            "uuid": "b84cdd53-ab30-4ed8-89fd-35edc78a5518"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Monday 20 January 2025  06:15:24 -0500 (0:00:00.379)       0:03:52.804 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.002351",
    "end": "2025-01-20 06:15:24.367016",
    "rc": 0,
    "start": "2025-01-20 06:15:24.364665"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518 /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Monday 20 January 2025  06:15:24 -0500 (0:00:00.393)       0:03:53.197 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002285",
    "end": "2025-01-20 06:15:24.736951",
    "failed_when_result": false,
    "rc": 0,
    "start": "2025-01-20 06:15:24.734666"
}

STDOUT:

luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518 /dev/sda1 -

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Monday 20 January 2025  06:15:24 -0500 (0:00:00.375)       0:03:53.573 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node1

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Monday 20 January 2025  06:15:24 -0500 (0:00:00.082)       0:03:53.655 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Monday 20 January 2025  06:15:24 -0500 (0:00:00.031)       0:03:53.687 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Monday 20 January 2025  06:15:24 -0500 (0:00:00.023)       0:03:53.711 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Monday 20 January 2025  06:15:24 -0500 (0:00:00.020)       0:03:53.731 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node1

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Monday 20 January 2025  06:15:25 -0500 (0:00:00.062)       0:03:53.794 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Monday 20 January 2025  06:15:25 -0500 (0:00:00.026)       0:03:53.820 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Monday 20 January 2025  06:15:25 -0500 (0:00:00.025)       0:03:53.846 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Monday 20 January 2025  06:15:25 -0500 (0:00:00.021)       0:03:53.868 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Monday 20 January 2025  06:15:25 -0500 (0:00:00.020)       0:03:53.889 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Monday 20 January 2025  06:15:25 -0500 (0:00:00.022)       0:03:53.911 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Monday 20 January 2025  06:15:25 -0500 (0:00:00.019)       0:03:53.930 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Monday 20 January 2025  06:15:25 -0500 (0:00:00.021)       0:03:53.951 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Monday 20 January 2025  06:15:25 -0500 (0:00:00.019)       0:03:53.971 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Monday 20 January 2025  06:15:25 -0500 (0:00:00.021)       0:03:53.992 ******** 
ok: [managed-node1] => {
    "changed": false,
    "failed_when_result": false,
    "rc": 1
}

STDERR:

Shared connection to 10.31.8.138 closed.



MSG:

non-zero return code

TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:74
Monday 20 January 2025  06:15:25 -0500 (0:00:00.400)       0:03:54.393 ******** 

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:84
Monday 20 January 2025  06:15:25 -0500 (0:00:00.020)       0:03:54.413 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node1

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Monday 20 January 2025  06:15:25 -0500 (0:00:00.041)       0:03:54.454 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Monday 20 January 2025  06:15:25 -0500 (0:00:00.020)       0:03:54.474 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Monday 20 January 2025  06:15:25 -0500 (0:00:00.019)       0:03:54.494 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Monday 20 January 2025  06:15:25 -0500 (0:00:00.019)       0:03:54.514 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Monday 20 January 2025  06:15:25 -0500 (0:00:00.021)       0:03:54.536 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Monday 20 January 2025  06:15:25 -0500 (0:00:00.021)       0:03:54.557 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Monday 20 January 2025  06:15:25 -0500 (0:00:00.022)       0:03:54.579 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Monday 20 January 2025  06:15:25 -0500 (0:00:00.020)       0:03:54.600 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Monday 20 January 2025  06:15:25 -0500 (0:00:00.021)       0:03:54.621 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Monday 20 January 2025  06:15:25 -0500 (0:00:00.021)       0:03:54.643 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Monday 20 January 2025  06:15:25 -0500 (0:00:00.023)       0:03:54.666 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:87
Monday 20 January 2025  06:15:25 -0500 (0:00:00.023)       0:03:54.690 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node1

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Monday 20 January 2025  06:15:25 -0500 (0:00:00.054)       0:03:54.744 ******** 
skipping: [managed-node1] => (item={'encryption': True, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': 'luks2', 'encryption_password': 'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518', '_raw_device': '/dev/sda1', '_mount_id': '/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/sda1'})  => {
    "ansible_loop_var": "storage_test_lvmraid_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_lvmraid_volume": {
        "_device": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
        "_kernel_device": "/dev/dm-0",
        "_mount_id": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
        "_raw_device": "/dev/sda1",
        "_raw_kernel_device": "/dev/sda1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [],
        "encryption": true,
        "encryption_cipher": null,
        "encryption_key": null,
        "encryption_key_size": null,
        "encryption_luks_version": "luks2",
        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "xfs",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_group": null,
        "mount_mode": null,
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "mount_user": null,
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "raid_stripe_size": null,
        "size": "4g",
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:90
Monday 20 January 2025  06:15:26 -0500 (0:00:00.030)       0:03:54.774 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node1

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Monday 20 January 2025  06:15:26 -0500 (0:00:00.041)       0:03:54.816 ******** 
skipping: [managed-node1] => (item={'encryption': True, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': 'luks2', 'encryption_password': 'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518', '_raw_device': '/dev/sda1', '_mount_id': '/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/sda1'})  => {
    "ansible_loop_var": "storage_test_thin_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_thin_volume": {
        "_device": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
        "_kernel_device": "/dev/dm-0",
        "_mount_id": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
        "_raw_device": "/dev/sda1",
        "_raw_kernel_device": "/dev/sda1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [],
        "encryption": true,
        "encryption_cipher": null,
        "encryption_key": null,
        "encryption_key_size": null,
        "encryption_luks_version": "luks2",
        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "xfs",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_group": null,
        "mount_mode": null,
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "mount_user": null,
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "raid_stripe_size": null,
        "size": "4g",
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:93
Monday 20 January 2025  06:15:26 -0500 (0:00:00.029)       0:03:54.845 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node1

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Monday 20 January 2025  06:15:26 -0500 (0:00:00.046)       0:03:54.892 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Monday 20 January 2025  06:15:26 -0500 (0:00:00.026)       0:03:54.919 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Monday 20 January 2025  06:15:26 -0500 (0:00:00.023)       0:03:54.942 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Monday 20 January 2025  06:15:26 -0500 (0:00:00.022)       0:03:54.965 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:96
Monday 20 January 2025  06:15:26 -0500 (0:00:00.022)       0:03:54.987 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node1

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Monday 20 January 2025  06:15:26 -0500 (0:00:00.046)       0:03:55.034 ******** 
skipping: [managed-node1] => (item={'encryption': True, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': 'luks2', 'encryption_password': 'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518', '_raw_device': '/dev/sda1', '_mount_id': '/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/sda1'})  => {
    "ansible_loop_var": "storage_test_vdo_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_vdo_volume": {
        "_device": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
        "_kernel_device": "/dev/dm-0",
        "_mount_id": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
        "_raw_device": "/dev/sda1",
        "_raw_kernel_device": "/dev/sda1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [],
        "encryption": true,
        "encryption_cipher": null,
        "encryption_key": null,
        "encryption_key_size": null,
        "encryption_luks_version": "luks2",
        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "xfs",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_group": null,
        "mount_mode": null,
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "mount_user": null,
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "raid_stripe_size": null,
        "size": "4g",
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:99
Monday 20 January 2025  06:15:26 -0500 (0:00:00.030)       0:03:55.064 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node1

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Monday 20 January 2025  06:15:26 -0500 (0:00:00.052)       0:03:55.117 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Monday 20 January 2025  06:15:26 -0500 (0:00:00.021)       0:03:55.139 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Monday 20 January 2025  06:15:26 -0500 (0:00:00.021)       0:03:55.161 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Monday 20 January 2025  06:15:26 -0500 (0:00:00.020)       0:03:55.182 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Monday 20 January 2025  06:15:26 -0500 (0:00:00.021)       0:03:55.203 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Monday 20 January 2025  06:15:26 -0500 (0:00:00.022)       0:03:55.225 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:102
Monday 20 January 2025  06:15:26 -0500 (0:00:00.020)       0:03:55.245 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Monday 20 January 2025  06:15:26 -0500 (0:00:00.019)       0:03:55.265 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node1

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Monday 20 January 2025  06:15:26 -0500 (0:00:00.037)       0:03:55.303 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Monday 20 January 2025  06:15:26 -0500 (0:00:00.032)       0:03:55.335 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node1

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Monday 20 January 2025  06:15:26 -0500 (0:00:00.156)       0:03:55.491 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Monday 20 January 2025  06:15:26 -0500 (0:00:00.036)       0:03:55.528 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Monday 20 January 2025  06:15:26 -0500 (0:00:00.038)       0:03:55.566 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Monday 20 January 2025  06:15:26 -0500 (0:00:00.033)       0:03:55.600 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Monday 20 January 2025  06:15:26 -0500 (0:00:00.089)       0:03:55.689 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Monday 20 January 2025  06:15:26 -0500 (0:00:00.035)       0:03:55.725 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Monday 20 January 2025  06:15:27 -0500 (0:00:00.043)       0:03:55.768 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Monday 20 January 2025  06:15:27 -0500 (0:00:00.043)       0:03:55.812 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Monday 20 January 2025  06:15:27 -0500 (0:00:00.033)       0:03:55.845 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Monday 20 January 2025  06:15:27 -0500 (0:00:00.032)       0:03:55.878 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Monday 20 January 2025  06:15:27 -0500 (0:00:00.033)       0:03:55.911 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Monday 20 January 2025  06:15:27 -0500 (0:00:00.032)       0:03:55.944 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Monday 20 January 2025  06:15:27 -0500 (0:00:00.063)       0:03:56.008 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Monday 20 January 2025  06:15:27 -0500 (0:00:00.039)       0:03:56.047 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Monday 20 January 2025  06:15:27 -0500 (0:00:00.035)       0:03:56.082 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Monday 20 January 2025  06:15:27 -0500 (0:00:00.029)       0:03:56.112 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Monday 20 January 2025  06:15:27 -0500 (0:00:00.044)       0:03:56.156 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Monday 20 January 2025  06:15:27 -0500 (0:00:00.031)       0:03:56.188 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Monday 20 January 2025  06:15:27 -0500 (0:00:00.044)       0:03:56.233 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Monday 20 January 2025  06:15:27 -0500 (0:00:00.045)       0:03:56.278 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371717.3098264,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1737371717.3098264,
        "dev": 6,
        "device_type": 2049,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 627029,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/blockdevice",
        "mode": "0660",
        "mtime": 1737371717.3098264,
        "nlink": 1,
        "path": "/dev/sda1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Monday 20 January 2025  06:15:28 -0500 (0:00:00.492)       0:03:56.771 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Monday 20 January 2025  06:15:28 -0500 (0:00:00.053)       0:03:56.824 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Monday 20 January 2025  06:15:28 -0500 (0:00:00.048)       0:03:56.873 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Monday 20 January 2025  06:15:28 -0500 (0:00:00.045)       0:03:56.918 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "st_volume_type": "partition"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Monday 20 January 2025  06:15:28 -0500 (0:00:00.034)       0:03:56.953 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Monday 20 January 2025  06:15:28 -0500 (0:00:00.030)       0:03:56.983 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Monday 20 January 2025  06:15:28 -0500 (0:00:00.036)       0:03:57.020 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371717.4438264,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1737371717.4438264,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 626164,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1737371717.4438264,
        "nlink": 1,
        "path": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Monday 20 January 2025  06:15:28 -0500 (0:00:00.411)       0:03:57.431 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Monday 20 January 2025  06:15:31 -0500 (0:00:02.905)       0:04:00.337 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cryptsetup",
        "luksDump",
        "/dev/sda1"
    ],
    "delta": "0:00:00.009902",
    "end": "2025-01-20 06:15:31.888799",
    "rc": 0,
    "start": "2025-01-20 06:15:31.878897"
}

STDOUT:

LUKS header information
Version:       	2
Epoch:         	3
Metadata area: 	16384 [bytes]
Keyslots area: 	16744448 [bytes]
UUID:          	b84cdd53-ab30-4ed8-89fd-35edc78a5518
Label:         	(no label)
Subsystem:     	(no subsystem)
Flags:       	(no flags)

Data segments:
  0: crypt
	offset: 16777216 [bytes]
	length: (whole device)
	cipher: aes-xts-plain64
	sector: 512 [bytes]

Keyslots:
  0: luks2
	Key:        512 bits
	Priority:   normal
	Cipher:     aes-xts-plain64
	Cipher key: 512 bits
	PBKDF:      argon2i
	Time cost:  4
	Memory:     792898
	Threads:    2
	Salt:       c7 1b c5 3b 58 85 91 fb 5f 2b 09 04 27 01 9e ee 
	            6e 1f 78 5f e6 a3 1d cd 86 c8 74 b0 35 33 69 ab 
	AF stripes: 4000
	AF hash:    sha256
	Area offset:32768 [bytes]
	Area length:258048 [bytes]
	Digest ID:  0
Tokens:
Digests:
  0: pbkdf2
	Hash:       sha256
	Iterations: 120029
	Salt:       9f 3a b9 fa 3b 92 bd 9d ee 44 bb 5b f7 2c 0f d3 
	            97 5c 4f cc 16 b3 a1 7e 27 68 ea 78 21 d0 ef 1a 
	Digest:     1d 02 ee a7 55 66 8c de 1a 38 cf f9 fa 93 be 47 
	            44 91 74 9b 6a f8 a6 77 f5 a0 6a 86 39 84 c2 1e 

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Monday 20 January 2025  06:15:31 -0500 (0:00:00.372)       0:04:00.709 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Monday 20 January 2025  06:15:31 -0500 (0:00:00.028)       0:04:00.737 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Monday 20 January 2025  06:15:32 -0500 (0:00:00.028)       0:04:00.766 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Monday 20 January 2025  06:15:32 -0500 (0:00:00.026)       0:04:00.792 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Monday 20 January 2025  06:15:32 -0500 (0:00:00.026)       0:04:00.818 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Monday 20 January 2025  06:15:32 -0500 (0:00:00.034)       0:04:00.853 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Monday 20 January 2025  06:15:32 -0500 (0:00:00.022)       0:04:00.875 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Monday 20 January 2025  06:15:32 -0500 (0:00:00.023)       0:04:00.898 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [
            "luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518 /dev/sda1 -"
        ],
        "_storage_test_expected_crypttab_entries": "1",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Monday 20 January 2025  06:15:32 -0500 (0:00:00.028)       0:04:00.927 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Monday 20 January 2025  06:15:32 -0500 (0:00:00.024)       0:04:00.951 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Monday 20 January 2025  06:15:32 -0500 (0:00:00.035)       0:04:00.987 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Monday 20 January 2025  06:15:32 -0500 (0:00:00.032)       0:04:01.019 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Monday 20 January 2025  06:15:32 -0500 (0:00:00.031)       0:04:01.050 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Monday 20 January 2025  06:15:32 -0500 (0:00:00.026)       0:04:01.077 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Monday 20 January 2025  06:15:32 -0500 (0:00:00.030)       0:04:01.107 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Monday 20 January 2025  06:15:32 -0500 (0:00:00.030)       0:04:01.137 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Monday 20 January 2025  06:15:32 -0500 (0:00:00.030)       0:04:01.167 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Monday 20 January 2025  06:15:32 -0500 (0:00:00.034)       0:04:01.202 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Monday 20 January 2025  06:15:32 -0500 (0:00:00.030)       0:04:01.233 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Monday 20 January 2025  06:15:32 -0500 (0:00:00.030)       0:04:01.264 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Monday 20 January 2025  06:15:32 -0500 (0:00:00.029)       0:04:01.293 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Monday 20 January 2025  06:15:32 -0500 (0:00:00.026)       0:04:01.320 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Monday 20 January 2025  06:15:32 -0500 (0:00:00.021)       0:04:01.342 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Monday 20 January 2025  06:15:32 -0500 (0:00:00.024)       0:04:01.367 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Monday 20 January 2025  06:15:32 -0500 (0:00:00.028)       0:04:01.395 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Monday 20 January 2025  06:15:32 -0500 (0:00:00.023)       0:04:01.419 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Monday 20 January 2025  06:15:32 -0500 (0:00:00.026)       0:04:01.445 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Monday 20 January 2025  06:15:32 -0500 (0:00:00.023)       0:04:01.468 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Monday 20 January 2025  06:15:32 -0500 (0:00:00.026)       0:04:01.495 ******** 
skipping: [managed-node1] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Monday 20 January 2025  06:15:32 -0500 (0:00:00.022)       0:04:01.518 ******** 
skipping: [managed-node1] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Monday 20 January 2025  06:15:32 -0500 (0:00:00.021)       0:04:01.539 ******** 
skipping: [managed-node1] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Monday 20 January 2025  06:15:32 -0500 (0:00:00.021)       0:04:01.560 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Monday 20 January 2025  06:15:32 -0500 (0:00:00.022)       0:04:01.583 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Monday 20 January 2025  06:15:32 -0500 (0:00:00.020)       0:04:01.603 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Monday 20 January 2025  06:15:32 -0500 (0:00:00.020)       0:04:01.623 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Monday 20 January 2025  06:15:32 -0500 (0:00:00.021)       0:04:01.645 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Monday 20 January 2025  06:15:32 -0500 (0:00:00.019)       0:04:01.664 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Monday 20 January 2025  06:15:32 -0500 (0:00:00.019)       0:04:01.684 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Monday 20 January 2025  06:15:32 -0500 (0:00:00.021)       0:04:01.705 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Monday 20 January 2025  06:15:32 -0500 (0:00:00.019)       0:04:01.725 ******** 
skipping: [managed-node1] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Monday 20 January 2025  06:15:32 -0500 (0:00:00.019)       0:04:01.744 ******** 
skipping: [managed-node1] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Monday 20 January 2025  06:15:33 -0500 (0:00:00.021)       0:04:01.765 ******** 
skipping: [managed-node1] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Monday 20 January 2025  06:15:33 -0500 (0:00:00.019)       0:04:01.785 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Monday 20 January 2025  06:15:33 -0500 (0:00:00.018)       0:04:01.804 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Monday 20 January 2025  06:15:33 -0500 (0:00:00.019)       0:04:01.824 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Monday 20 January 2025  06:15:33 -0500 (0:00:00.019)       0:04:01.843 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Monday 20 January 2025  06:15:33 -0500 (0:00:00.020)       0:04:01.863 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Monday 20 January 2025  06:15:33 -0500 (0:00:00.021)       0:04:01.884 ******** 
ok: [managed-node1] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Monday 20 January 2025  06:15:33 -0500 (0:00:00.022)       0:04:01.907 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Monday 20 January 2025  06:15:33 -0500 (0:00:00.022)       0:04:01.929 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Monday 20 January 2025  06:15:33 -0500 (0:00:00.021)       0:04:01.951 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Monday 20 January 2025  06:15:33 -0500 (0:00:00.019)       0:04:01.970 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Monday 20 January 2025  06:15:33 -0500 (0:00:00.022)       0:04:01.993 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Monday 20 January 2025  06:15:33 -0500 (0:00:00.060)       0:04:02.053 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Monday 20 January 2025  06:15:33 -0500 (0:00:00.022)       0:04:02.076 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Monday 20 January 2025  06:15:33 -0500 (0:00:00.020)       0:04:02.097 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Monday 20 January 2025  06:15:33 -0500 (0:00:00.021)       0:04:02.118 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Monday 20 January 2025  06:15:33 -0500 (0:00:00.021)       0:04:02.140 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Monday 20 January 2025  06:15:33 -0500 (0:00:00.019)       0:04:02.160 ******** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Monday 20 January 2025  06:15:33 -0500 (0:00:00.020)       0:04:02.180 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Create a file] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml:12
Monday 20 January 2025  06:15:33 -0500 (0:00:00.019)       0:04:02.199 ******** 
changed: [managed-node1] => {
    "changed": true,
    "dest": "/opt/test1/quux",
    "gid": 0,
    "group": "root",
    "mode": "0644",
    "owner": "root",
    "secontext": "unconfined_u:object_r:unlabeled_t:s0",
    "size": 0,
    "state": "file",
    "uid": 0
}

TASK [Test for correct handling of safe_mode] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:247
Monday 20 January 2025  06:15:33 -0500 (0:00:00.356)       0:04:02.556 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node1

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Monday 20 January 2025  06:15:33 -0500 (0:00:00.053)       0:04:02.610 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_pools_global": [],
        "storage_safe_mode_global": true,
        "storage_volumes_global": []
    },
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Monday 20 January 2025  06:15:33 -0500 (0:00:00.026)       0:04:02.637 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:15:33 -0500 (0:00:00.028)       0:04:02.665 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:15:33 -0500 (0:00:00.031)       0:04:02.696 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:15:33 -0500 (0:00:00.024)       0:04:02.721 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:15:34 -0500 (0:00:00.055)       0:04:02.777 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:15:34 -0500 (0:00:00.021)       0:04:02.798 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:15:34 -0500 (0:00:00.019)       0:04:02.818 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:15:34 -0500 (0:00:00.020)       0:04:02.839 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:15:34 -0500 (0:00:00.022)       0:04:02.861 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:15:34 -0500 (0:00:00.049)       0:04:02.911 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:15:37 -0500 (0:00:02.854)       0:04:05.765 ******** 
ok: [managed-node1] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "partition",
            "volumes": [
                {
                    "encryption": false,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "yabbadabbadoo",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g",
                    "type": "partition"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:15:37 -0500 (0:00:00.034)       0:04:05.800 ******** 
ok: [managed-node1] => {
    "storage_volumes": []
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:15:37 -0500 (0:00:00.032)       0:04:05.833 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:15:41 -0500 (0:00:04.359)       0:04:10.193 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:15:41 -0500 (0:00:00.087)       0:04:10.280 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:15:41 -0500 (0:00:00.033)       0:04:10.314 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:15:41 -0500 (0:00:00.033)       0:04:10.347 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:15:41 -0500 (0:00:00.031)       0:04:10.378 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:15:44 -0500 (0:00:02.922)       0:04:13.301 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-cryptsetup@luk...d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service": {
                "name": "systemd-cryptsetup@luk...d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
                "source": "systemd",
                "state": "inactive",
                "status": "generated"
            },
            "systemd-cryptsetup@luks\\x2d45ea6580\\x2d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service": {
                "name": "systemd-cryptsetup@luks\\x2d45ea6580\\x2d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:15:46 -0500 (0:00:01.751)       0:04:15.052 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": [
            "systemd-cryptsetup@luks\\x2d45ea6580\\x2d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
            "systemd-cryptsetup@luk...d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service"
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:15:46 -0500 (0:00:00.084)       0:04:15.137 ******** 
changed: [managed-node1] => (item=systemd-cryptsetup@luks\x2d45ea6580\x2d705d\x2d46a9\x2db81a\x2d78f69257cc6f.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luks\\x2d45ea6580\\x2d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
    "name": "systemd-cryptsetup@luks\\x2d45ea6580\\x2d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "After": "dev-sda.device systemd-journald.socket cryptsetup-pre.target system-systemd\\x2dcryptsetup.slice",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "Before": "umount.target cryptsetup.target",
        "BindsTo": "dev-sda.device",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "yes",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "Conflicts": "umount.target",
        "ControlPID": "0",
        "DefaultDependencies": "no",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "Cryptography Setup for luks-45ea6580-705d-46a9-b81a-78f69257cc6f",
        "DevicePolicy": "auto",
        "Documentation": "man:crypttab(5) man:systemd-cryptsetup-generator(8) man:systemd-cryptsetup@.service(8)",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-45ea6580-705d-46a9-b81a-78f69257cc6f /dev/sda -  ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
        "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-45ea6580-705d-46a9-b81a-78f69257cc6f ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2d45ea6580\\x2d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luks\\x2d45ea6580\\x2d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
        "IgnoreOnIsolate": "yes",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "shared",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "0",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "65536",
        "LimitMEMLOCKSoft": "65536",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "262144",
        "LimitNOFILESoft": "1024",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadState": "loaded",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luks\\x2d45ea6580\\x2d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
        "NeedDaemonReload": "yes",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "yes",
        "RemoveIPC": "no",
        "RequiredBy": "cryptsetup.target",
        "Requires": "system-systemd\\x2dcryptsetup.slice",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "Slice": "system-systemd\\x2dcryptsetup.slice",
        "SourcePath": "/etc/crypttab",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "journal",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestamp": "Mon 2025-01-20 06:15:21 EST",
        "StateChangeTimestampMonotonic": "2539252780",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "infinity",
        "TimeoutStopUSec": "infinity",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "Type": "oneshot",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFilePreset": "disabled",
        "UnitFileState": "generated",
        "UtmpMode": "init",
        "WantedBy": "dev-sda.device",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}
changed: [managed-node1] => (item=systemd-cryptsetup@luk...d705d\x2d46a9\x2db81a\x2d78f69257cc6f.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luk...d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
    "name": "systemd-cryptsetup@luk...d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luk...d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luk...d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.NoSuchUnit \"Unit systemd-cryptsetup@luk...d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service not found.\"",
        "LoadState": "not-found",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luk...d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:15:48 -0500 (0:00:01.638)       0:04:16.775 ******** 
fatal: [managed-node1]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

cannot remove existing formatting on device 'luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518' in safe mode due to encryption removal

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Monday 20 January 2025  06:15:52 -0500 (0:00:04.451)       0:04:21.227 ******** 
fatal: [managed-node1]: FAILED! => {
    "changed": false
}

MSG:

{'msg': "cannot remove existing formatting on device 'luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518' in safe mode due to encryption removal", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'encryption_clevis_pin': None, 'encryption_tang_url': None, 'encryption_tang_thumbprint': None, 'grow_to_fill': False, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'shared': False, 'state': 'present', 'type': 'partition', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': 0, 'encryption_luks_version': 'luks2', 'encryption_password': 'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}]}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'grow_to_fill': False, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, 'shared': False}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:15:52 -0500 (0:00:00.027)       0:04:21.254 ******** 
changed: [managed-node1] => (item=systemd-cryptsetup@luks\x2d45ea6580\x2d705d\x2d46a9\x2db81a\x2d78f69257cc6f.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luks\\x2d45ea6580\\x2d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
    "name": "systemd-cryptsetup@luks\\x2d45ea6580\\x2d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luks\\x2d45ea6580\\x2d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/dev/null",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luks\\x2d45ea6580\\x2d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luks\\x2d45ea6580\\x2d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service is masked.\"",
        "LoadState": "masked",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luks\\x2d45ea6580\\x2d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFileState": "masked",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}
changed: [managed-node1] => (item=systemd-cryptsetup@luk...d705d\x2d46a9\x2db81a\x2d78f69257cc6f.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luk...d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
    "name": "systemd-cryptsetup@luk...d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luk...d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/dev/null",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luk...d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luk...d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service is masked.\"",
        "LoadState": "masked",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luk...d705d\\x2d46a9\\x2db81a\\x2d78f69257cc6f.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFileState": "masked",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Monday 20 January 2025  06:15:53 -0500 (0:00:01.256)       0:04:22.511 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Monday 20 January 2025  06:15:53 -0500 (0:00:00.025)       0:04:22.537 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Monday 20 January 2025  06:15:53 -0500 (0:00:00.030)       0:04:22.567 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Stat the file] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:11
Monday 20 January 2025  06:15:53 -0500 (0:00:00.020)       0:04:22.588 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371733.7358558,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1737371733.7358558,
        "dev": 64768,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 131,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0644",
        "mtime": 1737371733.7358558,
        "nlink": 1,
        "path": "/opt/test1/quux",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "3349399803",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Assert file presence] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:16
Monday 20 January 2025  06:15:54 -0500 (0:00:00.381)       0:04:22.969 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Remove the encryption layer] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:272
Monday 20 January 2025  06:15:54 -0500 (0:00:00.036)       0:04:23.006 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:15:54 -0500 (0:00:00.132)       0:04:23.138 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:15:54 -0500 (0:00:00.046)       0:04:23.185 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:15:54 -0500 (0:00:00.037)       0:04:23.222 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:15:54 -0500 (0:00:00.085)       0:04:23.308 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:15:54 -0500 (0:00:00.032)       0:04:23.340 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:15:54 -0500 (0:00:00.031)       0:04:23.372 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:15:54 -0500 (0:00:00.030)       0:04:23.402 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:15:54 -0500 (0:00:00.031)       0:04:23.433 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:15:54 -0500 (0:00:00.122)       0:04:23.556 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:15:57 -0500 (0:00:02.917)       0:04:26.473 ******** 
ok: [managed-node1] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "partition",
            "volumes": [
                {
                    "encryption": false,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "yabbadabbadoo",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g",
                    "type": "partition"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:15:57 -0500 (0:00:00.042)       0:04:26.516 ******** 
ok: [managed-node1] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:15:57 -0500 (0:00:00.045)       0:04:26.561 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:16:02 -0500 (0:00:04.289)       0:04:30.850 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:16:02 -0500 (0:00:00.038)       0:04:30.889 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:16:02 -0500 (0:00:00.021)       0:04:30.910 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:16:02 -0500 (0:00:00.021)       0:04:30.932 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:16:02 -0500 (0:00:00.020)       0:04:30.953 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:16:05 -0500 (0:00:02.839)       0:04:33.792 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service": {
                "name": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
                "source": "systemd",
                "state": "inactive",
                "status": "generated"
            },
            "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service": {
                "name": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:16:06 -0500 (0:00:01.648)       0:04:35.441 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": [
            "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
            "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service"
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:16:06 -0500 (0:00:00.037)       0:04:35.478 ******** 
changed: [managed-node1] => (item=systemd-cryptsetup@luks\x2db84cdd53\x2dab30\x2d4ed8\x2d89fd\x2d35edc78a5518.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
    "name": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "After": "dev-sda1.device cryptsetup-pre.target systemd-journald.socket system-systemd\\x2dcryptsetup.slice",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "Before": "cryptsetup.target umount.target",
        "BindsTo": "dev-sda1.device",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "yes",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "Conflicts": "umount.target",
        "ControlPID": "0",
        "DefaultDependencies": "no",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "Cryptography Setup for luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
        "DevicePolicy": "auto",
        "Documentation": "man:crypttab(5) man:systemd-cryptsetup-generator(8) man:systemd-cryptsetup@.service(8)",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518 /dev/sda1 -  ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
        "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "IgnoreOnIsolate": "yes",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "shared",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "0",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "65536",
        "LimitMEMLOCKSoft": "65536",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "262144",
        "LimitNOFILESoft": "1024",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadState": "loaded",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "yes",
        "RemoveIPC": "no",
        "RequiredBy": "cryptsetup.target dev-mapper-luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.device",
        "Requires": "system-systemd\\x2dcryptsetup.slice",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "Slice": "system-systemd\\x2dcryptsetup.slice",
        "SourcePath": "/etc/crypttab",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "journal",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestamp": "Mon 2025-01-20 06:15:47 EST",
        "StateChangeTimestampMonotonic": "2565800548",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "infinity",
        "TimeoutStopUSec": "infinity",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "Type": "oneshot",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFilePreset": "disabled",
        "UnitFileState": "generated",
        "UtmpMode": "init",
        "WantedBy": "dev-sda1.device",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}
changed: [managed-node1] => (item=systemd-cryptsetup@luk...dab30\x2d4ed8\x2d89fd\x2d35edc78a5518.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
    "name": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.NoSuchUnit \"Unit systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service not found.\"",
        "LoadState": "not-found",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:16:08 -0500 (0:00:01.310)       0:04:36.789 ******** 
changed: [managed-node1] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
            "fs_type": "xfs"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/sda1",
            "fs_type": "luks"
        },
        {
            "action": "create format",
            "device": "/dev/sda1",
            "fs_type": "xfs"
        }
    ],
    "changed": true,
    "crypts": [
        {
            "backing_device": "/dev/sda1",
            "name": "luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
            "password": "-",
            "state": "absent"
        }
    ],
    "leaves": [
        "/dev/sda1",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1"
    ],
    "mounts": [
        {
            "fstype": "xfs",
            "path": "/opt/test1",
            "src": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
            "state": "absent"
        },
        {
            "dump": 0,
            "fstype": "xfs",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "UUID=d187465e-7b0f-4533-83d1-2db74230fc72",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "partition",
            "volumes": [
                {
                    "_device": "/dev/sda1",
                    "_kernel_device": "/dev/sda1",
                    "_mount_id": "UUID=d187465e-7b0f-4533-83d1-2db74230fc72",
                    "_raw_device": "/dev/sda1",
                    "_raw_kernel_device": "/dev/sda1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": 0,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "xfs",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "partition",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Monday 20 January 2025  06:16:12 -0500 (0:00:04.568)       0:04:41.358 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Monday 20 January 2025  06:16:12 -0500 (0:00:00.028)       0:04:41.386 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371720.478832,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "d34fe259c90f8846e9c8ce3434b7a9a72b05d286",
        "ctime": 1737371720.475832,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 207618244,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1737371720.475832,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1425,
        "uid": 0,
        "version": "261612921",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Monday 20 January 2025  06:16:13 -0500 (0:00:00.395)       0:04:41.781 ******** 
ok: [managed-node1] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:16:13 -0500 (0:00:00.385)       0:04:42.167 ******** 
changed: [managed-node1] => (item=systemd-cryptsetup@luks\x2db84cdd53\x2dab30\x2d4ed8\x2d89fd\x2d35edc78a5518.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
    "name": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/dev/null",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service is masked.\"",
        "LoadState": "masked",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "RequiredBy": "dev-mapper-luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.device cryptsetup.target",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestamp": "Mon 2025-01-20 06:15:47 EST",
        "StateChangeTimestampMonotonic": "2565800548",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFileState": "masked",
        "UtmpMode": "init",
        "WantedBy": "dev-sda1.device",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}
changed: [managed-node1] => (item=systemd-cryptsetup@luk...dab30\x2d4ed8\x2d89fd\x2d35edc78a5518.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
    "name": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/dev/null",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service is masked.\"",
        "LoadState": "masked",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFileState": "masked",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Monday 20 January 2025  06:16:14 -0500 (0:00:01.480)       0:04:43.647 ******** 
ok: [managed-node1] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
                "fs_type": "xfs"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/sda1",
                "fs_type": "luks"
            },
            {
                "action": "create format",
                "device": "/dev/sda1",
                "fs_type": "xfs"
            }
        ],
        "changed": true,
        "crypts": [
            {
                "backing_device": "/dev/sda1",
                "name": "luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
                "password": "-",
                "state": "absent"
            }
        ],
        "failed": false,
        "leaves": [
            "/dev/sda1",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1"
        ],
        "mounts": [
            {
                "fstype": "xfs",
                "path": "/opt/test1",
                "src": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
                "state": "absent"
            },
            {
                "dump": 0,
                "fstype": "xfs",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "UUID=d187465e-7b0f-4533-83d1-2db74230fc72",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "partition",
                "volumes": [
                    {
                        "_device": "/dev/sda1",
                        "_kernel_device": "/dev/sda1",
                        "_mount_id": "UUID=d187465e-7b0f-4533-83d1-2db74230fc72",
                        "_raw_device": "/dev/sda1",
                        "_raw_kernel_device": "/dev/sda1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": 0,
                        "encryption_luks_version": "luks2",
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "xfs",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "partition",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Monday 20 January 2025  06:16:14 -0500 (0:00:00.049)       0:04:43.696 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "partition",
                "volumes": [
                    {
                        "_device": "/dev/sda1",
                        "_kernel_device": "/dev/sda1",
                        "_mount_id": "UUID=d187465e-7b0f-4533-83d1-2db74230fc72",
                        "_raw_device": "/dev/sda1",
                        "_raw_kernel_device": "/dev/sda1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": 0,
                        "encryption_luks_version": "luks2",
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "xfs",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "partition",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Monday 20 January 2025  06:16:14 -0500 (0:00:00.049)       0:04:43.746 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Monday 20 January 2025  06:16:15 -0500 (0:00:00.067)       0:04:43.814 ******** 
changed: [managed-node1] => (item={'src': '/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "fstype": "xfs",
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Monday 20 January 2025  06:16:15 -0500 (0:00:00.488)       0:04:44.303 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Monday 20 January 2025  06:16:16 -0500 (0:00:00.731)       0:04:45.034 ******** 
changed: [managed-node1] => (item={'src': 'UUID=d187465e-7b0f-4533-83d1-2db74230fc72', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "UUID=d187465e-7b0f-4533-83d1-2db74230fc72",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "UUID=d187465e-7b0f-4533-83d1-2db74230fc72"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Monday 20 January 2025  06:16:16 -0500 (0:00:00.477)       0:04:45.512 ******** 
skipping: [managed-node1] => (item={'src': 'UUID=d187465e-7b0f-4533-83d1-2db74230fc72', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "UUID=d187465e-7b0f-4533-83d1-2db74230fc72",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Monday 20 January 2025  06:16:16 -0500 (0:00:00.074)       0:04:45.587 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Monday 20 January 2025  06:16:17 -0500 (0:00:00.873)       0:04:46.460 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371724.7358396,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "6c49f08dafaf6a76ef4d5ada3434f835dded959e",
        "ctime": 1737371722.6078358,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 29360388,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0600",
        "mtime": 1737371722.6068358,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 54,
        "uid": 0,
        "version": "1260335235",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Monday 20 January 2025  06:16:18 -0500 (0:00:00.490)       0:04:46.951 ******** 
changed: [managed-node1] => (item={'backing_device': '/dev/sda1', 'name': 'luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518', 'password': '-', 'state': 'absent'}) => {
    "ansible_loop_var": "entry",
    "backup": "",
    "changed": true,
    "entry": {
        "backing_device": "/dev/sda1",
        "name": "luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
        "password": "-",
        "state": "absent"
    },
    "found": 1
}

MSG:

1 line(s) removed

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Monday 20 January 2025  06:16:18 -0500 (0:00:00.611)       0:04:47.563 ******** 
ok: [managed-node1]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:290
Monday 20 January 2025  06:16:19 -0500 (0:00:01.009)       0:04:48.572 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node1

TASK [Print out pool information] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Monday 20 January 2025  06:16:19 -0500 (0:00:00.104)       0:04:48.676 ******** 
ok: [managed-node1] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "partition",
            "volumes": [
                {
                    "_device": "/dev/sda1",
                    "_kernel_device": "/dev/sda1",
                    "_mount_id": "UUID=d187465e-7b0f-4533-83d1-2db74230fc72",
                    "_raw_device": "/dev/sda1",
                    "_raw_kernel_device": "/dev/sda1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": 0,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "xfs",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "partition",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Monday 20 January 2025  06:16:19 -0500 (0:00:00.056)       0:04:48.732 ******** 
skipping: [managed-node1] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Monday 20 January 2025  06:16:19 -0500 (0:00:00.030)       0:04:48.762 ******** 
ok: [managed-node1] => {
    "changed": false,
    "info": {
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/sda1",
            "size": "10G",
            "type": "partition",
            "uuid": "d187465e-7b0f-4533-83d1-2db74230fc72"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Monday 20 January 2025  06:16:20 -0500 (0:00:00.368)       0:04:49.131 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.002638",
    "end": "2025-01-20 06:16:20.721952",
    "rc": 0,
    "start": "2025-01-20 06:16:20.719314"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
UUID=d187465e-7b0f-4533-83d1-2db74230fc72 /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Monday 20 January 2025  06:16:20 -0500 (0:00:00.431)       0:04:49.563 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002178",
    "end": "2025-01-20 06:16:21.093824",
    "failed_when_result": false,
    "rc": 0,
    "start": "2025-01-20 06:16:21.091646"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Monday 20 January 2025  06:16:21 -0500 (0:00:00.349)       0:04:49.912 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node1

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Monday 20 January 2025  06:16:21 -0500 (0:00:00.041)       0:04:49.954 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Monday 20 January 2025  06:16:21 -0500 (0:00:00.020)       0:04:49.974 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Monday 20 January 2025  06:16:21 -0500 (0:00:00.020)       0:04:49.994 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Monday 20 January 2025  06:16:21 -0500 (0:00:00.020)       0:04:50.015 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node1

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Monday 20 January 2025  06:16:21 -0500 (0:00:00.046)       0:04:50.061 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Monday 20 January 2025  06:16:21 -0500 (0:00:00.051)       0:04:50.113 ******** 

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Monday 20 January 2025  06:16:21 -0500 (0:00:00.020)       0:04:50.133 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Monday 20 January 2025  06:16:21 -0500 (0:00:00.020)       0:04:50.154 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Monday 20 January 2025  06:16:21 -0500 (0:00:00.019)       0:04:50.174 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Monday 20 January 2025  06:16:21 -0500 (0:00:00.019)       0:04:50.194 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Monday 20 January 2025  06:16:21 -0500 (0:00:00.020)       0:04:50.214 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Monday 20 January 2025  06:16:21 -0500 (0:00:00.021)       0:04:50.235 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Monday 20 January 2025  06:16:21 -0500 (0:00:00.020)       0:04:50.256 ******** 

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Monday 20 January 2025  06:16:21 -0500 (0:00:00.017)       0:04:50.274 ******** 
ok: [managed-node1] => {
    "changed": false,
    "failed_when_result": false,
    "rc": 1
}

STDERR:

Shared connection to 10.31.8.138 closed.



MSG:

non-zero return code

TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:74
Monday 20 January 2025  06:16:21 -0500 (0:00:00.403)       0:04:50.677 ******** 

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:84
Monday 20 January 2025  06:16:21 -0500 (0:00:00.019)       0:04:50.697 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node1

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Monday 20 January 2025  06:16:21 -0500 (0:00:00.041)       0:04:50.738 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Monday 20 January 2025  06:16:22 -0500 (0:00:00.028)       0:04:50.766 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Monday 20 January 2025  06:16:22 -0500 (0:00:00.024)       0:04:50.790 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Monday 20 January 2025  06:16:22 -0500 (0:00:00.020)       0:04:50.811 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Monday 20 January 2025  06:16:22 -0500 (0:00:00.019)       0:04:50.831 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Monday 20 January 2025  06:16:22 -0500 (0:00:00.021)       0:04:50.852 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Monday 20 January 2025  06:16:22 -0500 (0:00:00.021)       0:04:50.874 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Monday 20 January 2025  06:16:22 -0500 (0:00:00.021)       0:04:50.895 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Monday 20 January 2025  06:16:22 -0500 (0:00:00.021)       0:04:50.917 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Monday 20 January 2025  06:16:22 -0500 (0:00:00.021)       0:04:50.938 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Monday 20 January 2025  06:16:22 -0500 (0:00:00.020)       0:04:50.959 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:87
Monday 20 January 2025  06:16:22 -0500 (0:00:00.021)       0:04:50.981 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node1

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Monday 20 January 2025  06:16:22 -0500 (0:00:00.040)       0:04:51.021 ******** 
skipping: [managed-node1] => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': 0, 'encryption_luks_version': 'luks2', 'encryption_password': 'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/sda1', '_raw_device': '/dev/sda1', '_mount_id': 'UUID=d187465e-7b0f-4533-83d1-2db74230fc72', '_kernel_device': '/dev/sda1', '_raw_kernel_device': '/dev/sda1'})  => {
    "ansible_loop_var": "storage_test_lvmraid_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_lvmraid_volume": {
        "_device": "/dev/sda1",
        "_kernel_device": "/dev/sda1",
        "_mount_id": "UUID=d187465e-7b0f-4533-83d1-2db74230fc72",
        "_raw_device": "/dev/sda1",
        "_raw_kernel_device": "/dev/sda1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [
            "sda"
        ],
        "encryption": false,
        "encryption_cipher": null,
        "encryption_key": null,
        "encryption_key_size": 0,
        "encryption_luks_version": "luks2",
        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "xfs",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_group": null,
        "mount_mode": null,
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "mount_user": null,
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "raid_stripe_size": null,
        "size": "4g",
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:90
Monday 20 January 2025  06:16:22 -0500 (0:00:00.028)       0:04:51.049 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node1

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Monday 20 January 2025  06:16:22 -0500 (0:00:00.042)       0:04:51.091 ******** 
skipping: [managed-node1] => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': 0, 'encryption_luks_version': 'luks2', 'encryption_password': 'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/sda1', '_raw_device': '/dev/sda1', '_mount_id': 'UUID=d187465e-7b0f-4533-83d1-2db74230fc72', '_kernel_device': '/dev/sda1', '_raw_kernel_device': '/dev/sda1'})  => {
    "ansible_loop_var": "storage_test_thin_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_thin_volume": {
        "_device": "/dev/sda1",
        "_kernel_device": "/dev/sda1",
        "_mount_id": "UUID=d187465e-7b0f-4533-83d1-2db74230fc72",
        "_raw_device": "/dev/sda1",
        "_raw_kernel_device": "/dev/sda1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [
            "sda"
        ],
        "encryption": false,
        "encryption_cipher": null,
        "encryption_key": null,
        "encryption_key_size": 0,
        "encryption_luks_version": "luks2",
        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "xfs",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_group": null,
        "mount_mode": null,
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "mount_user": null,
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "raid_stripe_size": null,
        "size": "4g",
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:93
Monday 20 January 2025  06:16:22 -0500 (0:00:00.027)       0:04:51.119 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node1

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Monday 20 January 2025  06:16:22 -0500 (0:00:00.045)       0:04:51.164 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Monday 20 January 2025  06:16:22 -0500 (0:00:00.024)       0:04:51.189 ******** 

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Monday 20 January 2025  06:16:22 -0500 (0:00:00.018)       0:04:51.208 ******** 

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Monday 20 January 2025  06:16:22 -0500 (0:00:00.019)       0:04:51.227 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:96
Monday 20 January 2025  06:16:22 -0500 (0:00:00.020)       0:04:51.248 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node1

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Monday 20 January 2025  06:16:22 -0500 (0:00:00.046)       0:04:51.295 ******** 
skipping: [managed-node1] => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': 0, 'encryption_luks_version': 'luks2', 'encryption_password': 'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/sda1', '_raw_device': '/dev/sda1', '_mount_id': 'UUID=d187465e-7b0f-4533-83d1-2db74230fc72', '_kernel_device': '/dev/sda1', '_raw_kernel_device': '/dev/sda1'})  => {
    "ansible_loop_var": "storage_test_vdo_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_vdo_volume": {
        "_device": "/dev/sda1",
        "_kernel_device": "/dev/sda1",
        "_mount_id": "UUID=d187465e-7b0f-4533-83d1-2db74230fc72",
        "_raw_device": "/dev/sda1",
        "_raw_kernel_device": "/dev/sda1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [
            "sda"
        ],
        "encryption": false,
        "encryption_cipher": null,
        "encryption_key": null,
        "encryption_key_size": 0,
        "encryption_luks_version": "luks2",
        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "xfs",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_group": null,
        "mount_mode": null,
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "mount_user": null,
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "raid_stripe_size": null,
        "size": "4g",
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:99
Monday 20 January 2025  06:16:22 -0500 (0:00:00.028)       0:04:51.323 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node1

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Monday 20 January 2025  06:16:22 -0500 (0:00:00.049)       0:04:51.373 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Monday 20 January 2025  06:16:22 -0500 (0:00:00.020)       0:04:51.394 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Monday 20 January 2025  06:16:22 -0500 (0:00:00.019)       0:04:51.414 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Monday 20 January 2025  06:16:22 -0500 (0:00:00.019)       0:04:51.433 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Monday 20 January 2025  06:16:22 -0500 (0:00:00.019)       0:04:51.453 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Monday 20 January 2025  06:16:22 -0500 (0:00:00.020)       0:04:51.473 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:102
Monday 20 January 2025  06:16:22 -0500 (0:00:00.021)       0:04:51.494 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Monday 20 January 2025  06:16:22 -0500 (0:00:00.019)       0:04:51.513 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node1

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Monday 20 January 2025  06:16:22 -0500 (0:00:00.037)       0:04:51.551 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Monday 20 January 2025  06:16:22 -0500 (0:00:00.025)       0:04:51.576 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node1

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Monday 20 January 2025  06:16:22 -0500 (0:00:00.139)       0:04:51.716 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/sda1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Monday 20 January 2025  06:16:22 -0500 (0:00:00.025)       0:04:51.741 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Monday 20 January 2025  06:16:23 -0500 (0:00:00.027)       0:04:51.769 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Monday 20 January 2025  06:16:23 -0500 (0:00:00.031)       0:04:51.801 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Monday 20 January 2025  06:16:23 -0500 (0:00:00.027)       0:04:51.828 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Monday 20 January 2025  06:16:23 -0500 (0:00:00.021)       0:04:51.850 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Monday 20 January 2025  06:16:23 -0500 (0:00:00.019)       0:04:51.870 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Monday 20 January 2025  06:16:23 -0500 (0:00:00.021)       0:04:51.891 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Monday 20 January 2025  06:16:23 -0500 (0:00:00.020)       0:04:51.911 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Monday 20 January 2025  06:16:23 -0500 (0:00:00.019)       0:04:51.931 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Monday 20 January 2025  06:16:23 -0500 (0:00:00.019)       0:04:51.951 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Monday 20 January 2025  06:16:23 -0500 (0:00:00.022)       0:04:51.973 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "UUID=d187465e-7b0f-4533-83d1-2db74230fc72 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Monday 20 January 2025  06:16:23 -0500 (0:00:00.042)       0:04:52.016 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Monday 20 January 2025  06:16:23 -0500 (0:00:00.025)       0:04:52.041 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Monday 20 January 2025  06:16:23 -0500 (0:00:00.024)       0:04:52.065 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Monday 20 January 2025  06:16:23 -0500 (0:00:00.019)       0:04:52.085 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Monday 20 January 2025  06:16:23 -0500 (0:00:00.025)       0:04:52.110 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Monday 20 January 2025  06:16:23 -0500 (0:00:00.021)       0:04:52.132 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Monday 20 January 2025  06:16:23 -0500 (0:00:00.028)       0:04:52.161 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Monday 20 January 2025  06:16:23 -0500 (0:00:00.032)       0:04:52.193 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371772.4789252,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1737371772.4789252,
        "dev": 6,
        "device_type": 2049,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 643335,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/blockdevice",
        "mode": "0660",
        "mtime": 1737371772.4789252,
        "nlink": 1,
        "path": "/dev/sda1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Monday 20 January 2025  06:16:23 -0500 (0:00:00.375)       0:04:52.569 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Monday 20 January 2025  06:16:23 -0500 (0:00:00.043)       0:04:52.612 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Monday 20 January 2025  06:16:23 -0500 (0:00:00.033)       0:04:52.645 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Monday 20 January 2025  06:16:23 -0500 (0:00:00.039)       0:04:52.685 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "st_volume_type": "partition"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Monday 20 January 2025  06:16:23 -0500 (0:00:00.035)       0:04:52.720 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Monday 20 January 2025  06:16:23 -0500 (0:00:00.033)       0:04:52.754 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Monday 20 January 2025  06:16:24 -0500 (0:00:00.037)       0:04:52.792 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Monday 20 January 2025  06:16:24 -0500 (0:00:00.034)       0:04:52.826 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Monday 20 January 2025  06:16:27 -0500 (0:00:02.991)       0:04:55.818 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Monday 20 January 2025  06:16:27 -0500 (0:00:00.036)       0:04:55.855 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Monday 20 January 2025  06:16:27 -0500 (0:00:00.029)       0:04:55.885 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Monday 20 January 2025  06:16:27 -0500 (0:00:00.043)       0:04:55.928 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Monday 20 January 2025  06:16:27 -0500 (0:00:00.032)       0:04:55.961 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Monday 20 January 2025  06:16:27 -0500 (0:00:00.044)       0:04:56.006 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Monday 20 January 2025  06:16:27 -0500 (0:00:00.035)       0:04:56.041 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Monday 20 January 2025  06:16:27 -0500 (0:00:00.032)       0:04:56.074 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Monday 20 January 2025  06:16:27 -0500 (0:00:00.043)       0:04:56.118 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Monday 20 January 2025  06:16:27 -0500 (0:00:00.054)       0:04:56.172 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Monday 20 January 2025  06:16:27 -0500 (0:00:00.044)       0:04:56.217 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Monday 20 January 2025  06:16:27 -0500 (0:00:00.031)       0:04:56.248 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Monday 20 January 2025  06:16:27 -0500 (0:00:00.038)       0:04:56.287 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Monday 20 January 2025  06:16:27 -0500 (0:00:00.037)       0:04:56.325 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Monday 20 January 2025  06:16:27 -0500 (0:00:00.051)       0:04:56.376 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Monday 20 January 2025  06:16:27 -0500 (0:00:00.040)       0:04:56.416 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Monday 20 January 2025  06:16:27 -0500 (0:00:00.034)       0:04:56.451 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Monday 20 January 2025  06:16:27 -0500 (0:00:00.029)       0:04:56.480 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Monday 20 January 2025  06:16:27 -0500 (0:00:00.030)       0:04:56.511 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Monday 20 January 2025  06:16:27 -0500 (0:00:00.023)       0:04:56.535 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Monday 20 January 2025  06:16:27 -0500 (0:00:00.064)       0:04:56.600 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Monday 20 January 2025  06:16:27 -0500 (0:00:00.033)       0:04:56.634 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Monday 20 January 2025  06:16:27 -0500 (0:00:00.032)       0:04:56.666 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Monday 20 January 2025  06:16:27 -0500 (0:00:00.030)       0:04:56.696 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Monday 20 January 2025  06:16:27 -0500 (0:00:00.031)       0:04:56.728 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Monday 20 January 2025  06:16:27 -0500 (0:00:00.024)       0:04:56.752 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Monday 20 January 2025  06:16:28 -0500 (0:00:00.030)       0:04:56.782 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Monday 20 January 2025  06:16:28 -0500 (0:00:00.026)       0:04:56.809 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Monday 20 January 2025  06:16:28 -0500 (0:00:00.022)       0:04:56.831 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Monday 20 January 2025  06:16:28 -0500 (0:00:00.021)       0:04:56.853 ******** 
skipping: [managed-node1] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Monday 20 January 2025  06:16:28 -0500 (0:00:00.020)       0:04:56.874 ******** 
skipping: [managed-node1] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Monday 20 January 2025  06:16:28 -0500 (0:00:00.021)       0:04:56.895 ******** 
skipping: [managed-node1] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Monday 20 January 2025  06:16:28 -0500 (0:00:00.029)       0:04:56.925 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Monday 20 January 2025  06:16:28 -0500 (0:00:00.022)       0:04:56.947 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Monday 20 January 2025  06:16:28 -0500 (0:00:00.028)       0:04:56.976 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Monday 20 January 2025  06:16:28 -0500 (0:00:00.033)       0:04:57.010 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Monday 20 January 2025  06:16:28 -0500 (0:00:00.026)       0:04:57.036 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Monday 20 January 2025  06:16:28 -0500 (0:00:00.034)       0:04:57.071 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Monday 20 January 2025  06:16:28 -0500 (0:00:00.093)       0:04:57.165 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Monday 20 January 2025  06:16:28 -0500 (0:00:00.054)       0:04:57.219 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Monday 20 January 2025  06:16:28 -0500 (0:00:00.060)       0:04:57.279 ******** 
skipping: [managed-node1] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Monday 20 January 2025  06:16:28 -0500 (0:00:00.053)       0:04:57.333 ******** 
skipping: [managed-node1] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Monday 20 January 2025  06:16:28 -0500 (0:00:00.063)       0:04:57.397 ******** 
skipping: [managed-node1] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Monday 20 January 2025  06:16:28 -0500 (0:00:00.064)       0:04:57.462 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Monday 20 January 2025  06:16:28 -0500 (0:00:00.050)       0:04:57.512 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Monday 20 January 2025  06:16:28 -0500 (0:00:00.063)       0:04:57.576 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Monday 20 January 2025  06:16:28 -0500 (0:00:00.072)       0:04:57.648 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Monday 20 January 2025  06:16:28 -0500 (0:00:00.060)       0:04:57.709 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Monday 20 January 2025  06:16:29 -0500 (0:00:00.069)       0:04:57.778 ******** 
ok: [managed-node1] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Monday 20 January 2025  06:16:29 -0500 (0:00:00.064)       0:04:57.843 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Monday 20 January 2025  06:16:29 -0500 (0:00:00.061)       0:04:57.904 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Monday 20 January 2025  06:16:29 -0500 (0:00:00.035)       0:04:57.940 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Monday 20 January 2025  06:16:29 -0500 (0:00:00.031)       0:04:57.972 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Monday 20 January 2025  06:16:29 -0500 (0:00:00.034)       0:04:58.007 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Monday 20 January 2025  06:16:29 -0500 (0:00:00.048)       0:04:58.055 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Monday 20 January 2025  06:16:29 -0500 (0:00:00.032)       0:04:58.088 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Monday 20 January 2025  06:16:29 -0500 (0:00:00.034)       0:04:58.122 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Monday 20 January 2025  06:16:29 -0500 (0:00:00.036)       0:04:58.159 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Monday 20 January 2025  06:16:29 -0500 (0:00:00.053)       0:04:58.212 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Monday 20 January 2025  06:16:29 -0500 (0:00:00.037)       0:04:58.250 ******** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Monday 20 January 2025  06:16:29 -0500 (0:00:00.029)       0:04:58.279 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Create a file] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml:12
Monday 20 January 2025  06:16:29 -0500 (0:00:00.031)       0:04:58.310 ******** 
changed: [managed-node1] => {
    "changed": true,
    "dest": "/opt/test1/quux",
    "gid": 0,
    "group": "root",
    "mode": "0644",
    "owner": "root",
    "secontext": "unconfined_u:object_r:unlabeled_t:s0",
    "size": 0,
    "state": "file",
    "uid": 0
}

TASK [Test for correct handling of safe_mode] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:296
Monday 20 January 2025  06:16:29 -0500 (0:00:00.450)       0:04:58.761 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node1

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Monday 20 January 2025  06:16:30 -0500 (0:00:00.102)       0:04:58.863 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_pools_global": [],
        "storage_safe_mode_global": true,
        "storage_volumes_global": []
    },
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Monday 20 January 2025  06:16:30 -0500 (0:00:00.064)       0:04:58.928 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:16:30 -0500 (0:00:00.067)       0:04:58.995 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:16:30 -0500 (0:00:00.106)       0:04:59.102 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:16:30 -0500 (0:00:00.049)       0:04:59.151 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:16:30 -0500 (0:00:00.087)       0:04:59.238 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:16:30 -0500 (0:00:00.058)       0:04:59.296 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:16:30 -0500 (0:00:00.031)       0:04:59.328 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:16:30 -0500 (0:00:00.032)       0:04:59.360 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:16:30 -0500 (0:00:00.043)       0:04:59.404 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:16:30 -0500 (0:00:00.077)       0:04:59.481 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:16:33 -0500 (0:00:03.145)       0:05:02.626 ******** 
ok: [managed-node1] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "partition",
            "volumes": [
                {
                    "encryption": true,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "yabbadabbadoo",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g",
                    "type": "partition"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:16:33 -0500 (0:00:00.041)       0:05:02.668 ******** 
ok: [managed-node1] => {
    "storage_volumes": []
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:16:33 -0500 (0:00:00.043)       0:05:02.711 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "cryptsetup"
    ],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:16:37 -0500 (0:00:03.917)       0:05:06.629 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:16:37 -0500 (0:00:00.058)       0:05:06.687 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:16:37 -0500 (0:00:00.032)       0:05:06.719 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:16:37 -0500 (0:00:00.033)       0:05:06.752 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:16:38 -0500 (0:00:00.029)       0:05:06.782 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:16:40 -0500 (0:00:02.878)       0:05:09.661 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service": {
                "name": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
                "source": "systemd",
                "state": "inactive",
                "status": "generated"
            },
            "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service": {
                "name": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:16:42 -0500 (0:00:01.683)       0:05:11.345 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": [
            "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
            "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service"
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:16:42 -0500 (0:00:00.051)       0:05:11.396 ******** 
changed: [managed-node1] => (item=systemd-cryptsetup@luks\x2db84cdd53\x2dab30\x2d4ed8\x2d89fd\x2d35edc78a5518.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
    "name": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "After": "systemd-journald.socket dev-sda1.device cryptsetup-pre.target system-systemd\\x2dcryptsetup.slice",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "Before": "umount.target cryptsetup.target",
        "BindsTo": "dev-sda1.device",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "yes",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "Conflicts": "umount.target",
        "ControlPID": "0",
        "DefaultDependencies": "no",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "Cryptography Setup for luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518",
        "DevicePolicy": "auto",
        "Documentation": "man:crypttab(5) man:systemd-cryptsetup-generator(8) man:systemd-cryptsetup@.service(8)",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518 /dev/sda1 -  ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
        "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-b84cdd53-ab30-4ed8-89fd-35edc78a5518 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "IgnoreOnIsolate": "yes",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "shared",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "0",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "65536",
        "LimitMEMLOCKSoft": "65536",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "262144",
        "LimitNOFILESoft": "1024",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadState": "loaded",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "NeedDaemonReload": "yes",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "yes",
        "RemoveIPC": "no",
        "RequiredBy": "cryptsetup.target",
        "Requires": "system-systemd\\x2dcryptsetup.slice",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "Slice": "system-systemd\\x2dcryptsetup.slice",
        "SourcePath": "/etc/crypttab",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "journal",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestamp": "Mon 2025-01-20 06:15:47 EST",
        "StateChangeTimestampMonotonic": "2565800548",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "infinity",
        "TimeoutStopUSec": "infinity",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "Type": "oneshot",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFilePreset": "disabled",
        "UnitFileState": "generated",
        "UtmpMode": "init",
        "WantedBy": "dev-sda1.device",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}
changed: [managed-node1] => (item=systemd-cryptsetup@luk...dab30\x2d4ed8\x2d89fd\x2d35edc78a5518.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
    "name": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.NoSuchUnit \"Unit systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service not found.\"",
        "LoadState": "not-found",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:16:44 -0500 (0:00:01.396)       0:05:12.793 ******** 
fatal: [managed-node1]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

cannot remove existing formatting on device 'sda1' in safe mode due to adding encryption

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Monday 20 January 2025  06:16:48 -0500 (0:00:04.333)       0:05:17.126 ******** 
fatal: [managed-node1]: FAILED! => {
    "changed": false
}

MSG:

{'msg': "cannot remove existing formatting on device 'sda1' in safe mode due to adding encryption", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'encryption_clevis_pin': None, 'encryption_tang_url': None, 'encryption_tang_thumbprint': None, 'grow_to_fill': False, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'shared': False, 'state': 'present', 'type': 'partition', 'volumes': [{'encryption': True, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': 'luks2', 'encryption_password': 'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}]}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'grow_to_fill': False, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, 'shared': False}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:16:48 -0500 (0:00:00.043)       0:05:17.170 ******** 
changed: [managed-node1] => (item=systemd-cryptsetup@luks\x2db84cdd53\x2dab30\x2d4ed8\x2d89fd\x2d35edc78a5518.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
    "name": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/dev/null",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service is masked.\"",
        "LoadState": "masked",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luks\\x2db84cdd53\\x2dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFileState": "masked",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}
changed: [managed-node1] => (item=systemd-cryptsetup@luk...dab30\x2d4ed8\x2d89fd\x2d35edc78a5518.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
    "name": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/dev/null",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service is masked.\"",
        "LoadState": "masked",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luk...dab30\\x2d4ed8\\x2d89fd\\x2d35edc78a5518.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFileState": "masked",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Monday 20 January 2025  06:16:49 -0500 (0:00:01.345)       0:05:18.516 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Monday 20 January 2025  06:16:49 -0500 (0:00:00.025)       0:05:18.541 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Monday 20 January 2025  06:16:49 -0500 (0:00:00.030)       0:05:18.572 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Stat the file] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:11
Monday 20 January 2025  06:16:49 -0500 (0:00:00.020)       0:05:18.592 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371789.8929563,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1737371789.8929563,
        "dev": 2049,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 131,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0644",
        "mtime": 1737371789.8929563,
        "nlink": 1,
        "path": "/opt/test1/quux",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "2094954857",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Assert file presence] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:16
Monday 20 January 2025  06:16:50 -0500 (0:00:00.353)       0:05:18.945 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Create a key file] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:323
Monday 20 January 2025  06:16:50 -0500 (0:00:00.023)       0:05:18.969 ******** 
ok: [managed-node1] => {
    "changed": false,
    "gid": 0,
    "group": "root",
    "mode": "0600",
    "owner": "root",
    "path": "/tmp/storage_test1_y4m_lvlukskey",
    "secontext": "unconfined_u:object_r:user_tmp_t:s0",
    "size": 0,
    "state": "file",
    "uid": 0
}

TASK [Write the key into the key file] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:330
Monday 20 January 2025  06:16:50 -0500 (0:00:00.461)       0:05:19.431 ******** 
ok: [managed-node1] => {
    "changed": false,
    "checksum": "7a4dff3752e2baf5617c57eaac048e2b95e8af91",
    "dest": "/tmp/storage_test1_y4m_lvlukskey",
    "gid": 0,
    "group": "root",
    "md5sum": "4ac07b967150835c00d0865161e48744",
    "mode": "0600",
    "owner": "root",
    "secontext": "unconfined_u:object_r:user_tmp_t:s0",
    "size": 32,
    "src": "/root/.ansible/tmp/ansible-tmp-1737371810.7080345-143435-277900666304871/source",
    "state": "file",
    "uid": 0
}

TASK [Add encryption to the volume] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:337
Monday 20 January 2025  06:16:51 -0500 (0:00:00.798)       0:05:20.229 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:16:51 -0500 (0:00:00.050)       0:05:20.280 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:16:51 -0500 (0:00:00.050)       0:05:20.330 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:16:51 -0500 (0:00:00.040)       0:05:20.371 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:16:51 -0500 (0:00:00.079)       0:05:20.450 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:16:51 -0500 (0:00:00.033)       0:05:20.483 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:16:51 -0500 (0:00:00.029)       0:05:20.513 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:16:51 -0500 (0:00:00.030)       0:05:20.544 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:16:51 -0500 (0:00:00.026)       0:05:20.570 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:16:51 -0500 (0:00:00.072)       0:05:20.643 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:16:54 -0500 (0:00:02.936)       0:05:23.580 ******** 
ok: [managed-node1] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "partition",
            "volumes": [
                {
                    "encryption": true,
                    "encryption_key": "/tmp/storage_test1_y4m_lvlukskey",
                    "encryption_luks_version": "luks2",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g",
                    "type": "partition"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:16:54 -0500 (0:00:00.037)       0:05:23.617 ******** 
ok: [managed-node1] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:16:54 -0500 (0:00:00.040)       0:05:23.657 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "cryptsetup"
    ],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:16:58 -0500 (0:00:03.840)       0:05:27.498 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:16:58 -0500 (0:00:00.037)       0:05:27.535 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:16:58 -0500 (0:00:00.020)       0:05:27.555 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:16:58 -0500 (0:00:00.020)       0:05:27.575 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:16:58 -0500 (0:00:00.019)       0:05:27.594 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:17:01 -0500 (0:00:02.829)       0:05:30.424 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:17:03 -0500 (0:00:01.663)       0:05:32.087 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:17:03 -0500 (0:00:00.033)       0:05:32.120 ******** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:17:03 -0500 (0:00:00.019)       0:05:32.140 ******** 
changed: [managed-node1] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/sda1",
            "fs_type": "xfs"
        },
        {
            "action": "create format",
            "device": "/dev/sda1",
            "fs_type": "luks"
        },
        {
            "action": "create device",
            "device": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
            "fs_type": "xfs"
        }
    ],
    "changed": true,
    "crypts": [
        {
            "backing_device": "/dev/sda1",
            "name": "luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
            "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
            "state": "present"
        }
    ],
    "leaves": [
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1",
        "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81"
    ],
    "mounts": [
        {
            "fstype": "xfs",
            "path": "/opt/test1",
            "src": "UUID=d187465e-7b0f-4533-83d1-2db74230fc72",
            "state": "absent"
        },
        {
            "dump": 0,
            "fstype": "xfs",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "cryptsetup"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "partition",
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
                    "_raw_device": "/dev/sda1",
                    "_raw_kernel_device": "/dev/sda1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": true,
                    "encryption_cipher": null,
                    "encryption_key": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                    "encryption_key_size": null,
                    "encryption_luks_version": "luks2",
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "xfs",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "partition",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Monday 20 January 2025  06:17:18 -0500 (0:00:14.711)       0:05:46.851 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Monday 20 January 2025  06:17:18 -0500 (0:00:00.080)       0:05:46.932 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371776.6199324,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "95ec999355e6aea87f954216ebb515206f12c68c",
        "ctime": 1737371776.6169326,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 207618244,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1737371776.6169326,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1413,
        "uid": 0,
        "version": "261612921",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Monday 20 January 2025  06:17:18 -0500 (0:00:00.407)       0:05:47.340 ******** 
ok: [managed-node1] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:17:18 -0500 (0:00:00.407)       0:05:47.747 ******** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Monday 20 January 2025  06:17:19 -0500 (0:00:00.039)       0:05:47.787 ******** 
ok: [managed-node1] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/sda1",
                "fs_type": "xfs"
            },
            {
                "action": "create format",
                "device": "/dev/sda1",
                "fs_type": "luks"
            },
            {
                "action": "create device",
                "device": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
                "fs_type": "xfs"
            }
        ],
        "changed": true,
        "crypts": [
            {
                "backing_device": "/dev/sda1",
                "name": "luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
                "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                "state": "present"
            }
        ],
        "failed": false,
        "leaves": [
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1",
            "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81"
        ],
        "mounts": [
            {
                "fstype": "xfs",
                "path": "/opt/test1",
                "src": "UUID=d187465e-7b0f-4533-83d1-2db74230fc72",
                "state": "absent"
            },
            {
                "dump": 0,
                "fstype": "xfs",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "cryptsetup"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "partition",
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
                        "_raw_device": "/dev/sda1",
                        "_raw_kernel_device": "/dev/sda1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": true,
                        "encryption_cipher": null,
                        "encryption_key": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                        "encryption_key_size": null,
                        "encryption_luks_version": "luks2",
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "xfs",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "partition",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Monday 20 January 2025  06:17:19 -0500 (0:00:00.044)       0:05:47.831 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "partition",
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
                        "_raw_device": "/dev/sda1",
                        "_raw_kernel_device": "/dev/sda1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": true,
                        "encryption_cipher": null,
                        "encryption_key": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                        "encryption_key_size": null,
                        "encryption_luks_version": "luks2",
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "xfs",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "partition",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Monday 20 January 2025  06:17:19 -0500 (0:00:00.051)       0:05:47.883 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Monday 20 January 2025  06:17:19 -0500 (0:00:00.034)       0:05:47.917 ******** 
changed: [managed-node1] => (item={'src': 'UUID=d187465e-7b0f-4533-83d1-2db74230fc72', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "fstype": "xfs",
        "path": "/opt/test1",
        "src": "UUID=d187465e-7b0f-4533-83d1-2db74230fc72",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "UUID=d187465e-7b0f-4533-83d1-2db74230fc72"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Monday 20 January 2025  06:17:19 -0500 (0:00:00.415)       0:05:48.333 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Monday 20 January 2025  06:17:20 -0500 (0:00:00.838)       0:05:49.172 ******** 
changed: [managed-node1] => (item={'src': '/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Monday 20 January 2025  06:17:21 -0500 (0:00:00.748)       0:05:49.920 ******** 
skipping: [managed-node1] => (item={'src': '/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Monday 20 January 2025  06:17:21 -0500 (0:00:00.044)       0:05:49.965 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Monday 20 January 2025  06:17:22 -0500 (0:00:00.854)       0:05:50.820 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371781.0919406,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1737371778.6959362,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 184549593,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1737371778.6949363,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "3124594605",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Monday 20 January 2025  06:17:22 -0500 (0:00:00.433)       0:05:51.253 ******** 
changed: [managed-node1] => (item={'backing_device': '/dev/sda1', 'name': 'luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81', 'password': 'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', 'state': 'present'}) => {
    "ansible_loop_var": "entry",
    "backup": "",
    "changed": true,
    "entry": {
        "backing_device": "/dev/sda1",
        "name": "luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
        "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
        "state": "present"
    }
}

MSG:

line added

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Monday 20 January 2025  06:17:22 -0500 (0:00:00.506)       0:05:51.760 ******** 
ok: [managed-node1]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:355
Monday 20 January 2025  06:17:23 -0500 (0:00:00.973)       0:05:52.733 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node1

TASK [Print out pool information] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Monday 20 January 2025  06:17:24 -0500 (0:00:00.051)       0:05:52.785 ******** 
ok: [managed-node1] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "partition",
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
                    "_raw_device": "/dev/sda1",
                    "_raw_kernel_device": "/dev/sda1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": true,
                    "encryption_cipher": null,
                    "encryption_key": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                    "encryption_key_size": null,
                    "encryption_luks_version": "luks2",
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "xfs",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "partition",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Monday 20 January 2025  06:17:24 -0500 (0:00:00.046)       0:05:52.831 ******** 
skipping: [managed-node1] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Monday 20 January 2025  06:17:24 -0500 (0:00:00.032)       0:05:52.864 ******** 
ok: [managed-node1] => {
    "changed": false,
    "info": {
        "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
            "size": "10G",
            "type": "crypt",
            "uuid": "ef58b90d-e44b-44f5-bd96-f5612259c09f"
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda1": {
            "fstype": "crypto_LUKS",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda1",
            "size": "10G",
            "type": "partition",
            "uuid": "4fd96f37-c9c1-41aa-bce3-48eefe3c9f81"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Monday 20 January 2025  06:17:24 -0500 (0:00:00.474)       0:05:53.338 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.002639",
    "end": "2025-01-20 06:17:25.071217",
    "rc": 0,
    "start": "2025-01-20 06:17:25.068578"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81 /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Monday 20 January 2025  06:17:25 -0500 (0:00:00.574)       0:05:53.913 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002545",
    "end": "2025-01-20 06:17:25.485179",
    "failed_when_result": false,
    "rc": 0,
    "start": "2025-01-20 06:17:25.482634"
}

STDOUT:

luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81 /dev/sda1 VALUE_SPECIFIED_IN_NO_LOG_PARAMETER

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Monday 20 January 2025  06:17:25 -0500 (0:00:00.392)       0:05:54.305 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node1

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Monday 20 January 2025  06:17:25 -0500 (0:00:00.041)       0:05:54.346 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Monday 20 January 2025  06:17:25 -0500 (0:00:00.021)       0:05:54.368 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Monday 20 January 2025  06:17:25 -0500 (0:00:00.025)       0:05:54.394 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Monday 20 January 2025  06:17:25 -0500 (0:00:00.027)       0:05:54.422 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node1

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Monday 20 January 2025  06:17:25 -0500 (0:00:00.070)       0:05:54.492 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Monday 20 January 2025  06:17:25 -0500 (0:00:00.035)       0:05:54.527 ******** 

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Monday 20 January 2025  06:17:25 -0500 (0:00:00.035)       0:05:54.563 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Monday 20 January 2025  06:17:25 -0500 (0:00:00.031)       0:05:54.594 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Monday 20 January 2025  06:17:25 -0500 (0:00:00.032)       0:05:54.626 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Monday 20 January 2025  06:17:25 -0500 (0:00:00.033)       0:05:54.659 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Monday 20 January 2025  06:17:25 -0500 (0:00:00.034)       0:05:54.694 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Monday 20 January 2025  06:17:25 -0500 (0:00:00.038)       0:05:54.733 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Monday 20 January 2025  06:17:26 -0500 (0:00:00.081)       0:05:54.814 ******** 

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Monday 20 January 2025  06:17:26 -0500 (0:00:00.032)       0:05:54.847 ******** 
ok: [managed-node1] => {
    "changed": false,
    "failed_when_result": false,
    "rc": 1
}

STDERR:

Shared connection to 10.31.8.138 closed.



MSG:

non-zero return code

TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:74
Monday 20 January 2025  06:17:26 -0500 (0:00:00.472)       0:05:55.319 ******** 

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:84
Monday 20 January 2025  06:17:26 -0500 (0:00:00.038)       0:05:55.358 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node1

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Monday 20 January 2025  06:17:26 -0500 (0:00:00.063)       0:05:55.421 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Monday 20 January 2025  06:17:26 -0500 (0:00:00.034)       0:05:55.455 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Monday 20 January 2025  06:17:26 -0500 (0:00:00.031)       0:05:55.486 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Monday 20 January 2025  06:17:26 -0500 (0:00:00.032)       0:05:55.519 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Monday 20 January 2025  06:17:26 -0500 (0:00:00.035)       0:05:55.555 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Monday 20 January 2025  06:17:26 -0500 (0:00:00.032)       0:05:55.587 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Monday 20 January 2025  06:17:26 -0500 (0:00:00.032)       0:05:55.619 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Monday 20 January 2025  06:17:26 -0500 (0:00:00.035)       0:05:55.654 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Monday 20 January 2025  06:17:26 -0500 (0:00:00.030)       0:05:55.685 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Monday 20 January 2025  06:17:26 -0500 (0:00:00.031)       0:05:55.716 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Monday 20 January 2025  06:17:26 -0500 (0:00:00.036)       0:05:55.754 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:87
Monday 20 January 2025  06:17:27 -0500 (0:00:00.030)       0:05:55.784 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node1

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Monday 20 January 2025  06:17:27 -0500 (0:00:00.063)       0:05:55.847 ******** 
skipping: [managed-node1] => (item={'encryption': True, 'encryption_cipher': None, 'encryption_key': 'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', 'encryption_key_size': None, 'encryption_luks_version': 'luks2', 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81', '_raw_device': '/dev/sda1', '_mount_id': '/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/sda1'})  => {
    "ansible_loop_var": "storage_test_lvmraid_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_lvmraid_volume": {
        "_device": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
        "_kernel_device": "/dev/dm-0",
        "_mount_id": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
        "_raw_device": "/dev/sda1",
        "_raw_kernel_device": "/dev/sda1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [
            "sda"
        ],
        "encryption": true,
        "encryption_cipher": null,
        "encryption_key": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
        "encryption_key_size": null,
        "encryption_luks_version": "luks2",
        "encryption_password": null,
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "xfs",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_group": null,
        "mount_mode": null,
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "mount_user": null,
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "raid_stripe_size": null,
        "size": "4g",
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:90
Monday 20 January 2025  06:17:27 -0500 (0:00:00.042)       0:05:55.889 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node1

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Monday 20 January 2025  06:17:27 -0500 (0:00:00.066)       0:05:55.955 ******** 
skipping: [managed-node1] => (item={'encryption': True, 'encryption_cipher': None, 'encryption_key': 'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', 'encryption_key_size': None, 'encryption_luks_version': 'luks2', 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81', '_raw_device': '/dev/sda1', '_mount_id': '/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/sda1'})  => {
    "ansible_loop_var": "storage_test_thin_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_thin_volume": {
        "_device": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
        "_kernel_device": "/dev/dm-0",
        "_mount_id": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
        "_raw_device": "/dev/sda1",
        "_raw_kernel_device": "/dev/sda1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [
            "sda"
        ],
        "encryption": true,
        "encryption_cipher": null,
        "encryption_key": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
        "encryption_key_size": null,
        "encryption_luks_version": "luks2",
        "encryption_password": null,
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "xfs",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_group": null,
        "mount_mode": null,
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "mount_user": null,
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "raid_stripe_size": null,
        "size": "4g",
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:93
Monday 20 January 2025  06:17:27 -0500 (0:00:00.048)       0:05:56.004 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node1

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Monday 20 January 2025  06:17:27 -0500 (0:00:00.072)       0:05:56.076 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Monday 20 January 2025  06:17:27 -0500 (0:00:00.038)       0:05:56.115 ******** 

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Monday 20 January 2025  06:17:27 -0500 (0:00:00.027)       0:05:56.142 ******** 

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Monday 20 January 2025  06:17:27 -0500 (0:00:00.029)       0:05:56.172 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:96
Monday 20 January 2025  06:17:27 -0500 (0:00:00.028)       0:05:56.200 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node1

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Monday 20 January 2025  06:17:27 -0500 (0:00:00.068)       0:05:56.269 ******** 
skipping: [managed-node1] => (item={'encryption': True, 'encryption_cipher': None, 'encryption_key': 'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', 'encryption_key_size': None, 'encryption_luks_version': 'luks2', 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81', '_raw_device': '/dev/sda1', '_mount_id': '/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/sda1'})  => {
    "ansible_loop_var": "storage_test_vdo_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_vdo_volume": {
        "_device": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
        "_kernel_device": "/dev/dm-0",
        "_mount_id": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
        "_raw_device": "/dev/sda1",
        "_raw_kernel_device": "/dev/sda1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [
            "sda"
        ],
        "encryption": true,
        "encryption_cipher": null,
        "encryption_key": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
        "encryption_key_size": null,
        "encryption_luks_version": "luks2",
        "encryption_password": null,
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "xfs",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_group": null,
        "mount_mode": null,
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "mount_user": null,
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "raid_stripe_size": null,
        "size": "4g",
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:99
Monday 20 January 2025  06:17:27 -0500 (0:00:00.063)       0:05:56.332 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node1

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Monday 20 January 2025  06:17:27 -0500 (0:00:00.083)       0:05:56.416 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Monday 20 January 2025  06:17:27 -0500 (0:00:00.034)       0:05:56.450 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Monday 20 January 2025  06:17:27 -0500 (0:00:00.032)       0:05:56.482 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Monday 20 January 2025  06:17:27 -0500 (0:00:00.033)       0:05:56.516 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Monday 20 January 2025  06:17:27 -0500 (0:00:00.033)       0:05:56.549 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Monday 20 January 2025  06:17:27 -0500 (0:00:00.031)       0:05:56.580 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:102
Monday 20 January 2025  06:17:27 -0500 (0:00:00.031)       0:05:56.612 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Monday 20 January 2025  06:17:27 -0500 (0:00:00.025)       0:05:56.637 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node1

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Monday 20 January 2025  06:17:27 -0500 (0:00:00.058)       0:05:56.696 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Monday 20 January 2025  06:17:27 -0500 (0:00:00.043)       0:05:56.739 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node1

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Monday 20 January 2025  06:17:28 -0500 (0:00:00.194)       0:05:56.934 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Monday 20 January 2025  06:17:28 -0500 (0:00:00.041)       0:05:56.975 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Monday 20 January 2025  06:17:28 -0500 (0:00:00.044)       0:05:57.020 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Monday 20 January 2025  06:17:28 -0500 (0:00:00.089)       0:05:57.109 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Monday 20 January 2025  06:17:28 -0500 (0:00:00.041)       0:05:57.151 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Monday 20 January 2025  06:17:28 -0500 (0:00:00.032)       0:05:57.183 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Monday 20 January 2025  06:17:28 -0500 (0:00:00.029)       0:05:57.213 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Monday 20 January 2025  06:17:28 -0500 (0:00:00.031)       0:05:57.244 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Monday 20 January 2025  06:17:28 -0500 (0:00:00.031)       0:05:57.276 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Monday 20 January 2025  06:17:28 -0500 (0:00:00.033)       0:05:57.310 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Monday 20 January 2025  06:17:28 -0500 (0:00:00.048)       0:05:57.358 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Monday 20 January 2025  06:17:28 -0500 (0:00:00.031)       0:05:57.390 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Monday 20 January 2025  06:17:28 -0500 (0:00:00.064)       0:05:57.454 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Monday 20 January 2025  06:17:28 -0500 (0:00:00.041)       0:05:57.496 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Monday 20 January 2025  06:17:28 -0500 (0:00:00.038)       0:05:57.535 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Monday 20 January 2025  06:17:28 -0500 (0:00:00.035)       0:05:57.570 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Monday 20 January 2025  06:17:28 -0500 (0:00:00.040)       0:05:57.611 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Monday 20 January 2025  06:17:28 -0500 (0:00:00.028)       0:05:57.640 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Monday 20 January 2025  06:17:28 -0500 (0:00:00.045)       0:05:57.685 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Monday 20 January 2025  06:17:28 -0500 (0:00:00.045)       0:05:57.730 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371837.8100421,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1737371837.8100421,
        "dev": 6,
        "device_type": 2049,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 643335,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/blockdevice",
        "mode": "0660",
        "mtime": 1737371837.8100421,
        "nlink": 1,
        "path": "/dev/sda1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Monday 20 January 2025  06:17:29 -0500 (0:00:00.406)       0:05:58.137 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Monday 20 January 2025  06:17:29 -0500 (0:00:00.042)       0:05:58.179 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Monday 20 January 2025  06:17:29 -0500 (0:00:00.030)       0:05:58.210 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Monday 20 January 2025  06:17:29 -0500 (0:00:00.038)       0:05:58.248 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "st_volume_type": "partition"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Monday 20 January 2025  06:17:29 -0500 (0:00:00.036)       0:05:58.285 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Monday 20 January 2025  06:17:29 -0500 (0:00:00.026)       0:05:58.311 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Monday 20 January 2025  06:17:29 -0500 (0:00:00.029)       0:05:58.341 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371837.9570425,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1737371837.9570425,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 658487,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1737371837.9570425,
        "nlink": 1,
        "path": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Monday 20 January 2025  06:17:29 -0500 (0:00:00.395)       0:05:58.737 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Monday 20 January 2025  06:17:32 -0500 (0:00:02.900)       0:06:01.637 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cryptsetup",
        "luksDump",
        "/dev/sda1"
    ],
    "delta": "0:00:00.011165",
    "end": "2025-01-20 06:17:33.234483",
    "rc": 0,
    "start": "2025-01-20 06:17:33.223318"
}

STDOUT:

LUKS header information
Version:       	2
Epoch:         	3
Metadata area: 	16384 [bytes]
Keyslots area: 	16744448 [bytes]
UUID:          	4fd96f37-c9c1-41aa-bce3-48eefe3c9f81
Label:         	(no label)
Subsystem:     	(no subsystem)
Flags:       	(no flags)

Data segments:
  0: crypt
	offset: 16777216 [bytes]
	length: (whole device)
	cipher: aes-xts-plain64
	sector: 512 [bytes]

Keyslots:
  0: luks2
	Key:        512 bits
	Priority:   normal
	Cipher:     aes-xts-plain64
	Cipher key: 512 bits
	PBKDF:      argon2i
	Time cost:  4
	Memory:     793490
	Threads:    2
	Salt:       a2 24 98 52 ee 51 8d 1e f9 7d c2 c3 59 cf 91 24 
	            7f 32 d3 3b ba d9 cf 4c 8a a3 5d a5 87 f8 7d 2c 
	AF stripes: 4000
	AF hash:    sha256
	Area offset:32768 [bytes]
	Area length:258048 [bytes]
	Digest ID:  0
Tokens:
Digests:
  0: pbkdf2
	Hash:       sha256
	Iterations: 112798
	Salt:       9f 2a 18 09 a5 51 bd cb e8 9a 13 1f cf e6 8d 2f 
	            06 8c 03 7d 04 04 5e a2 d7 5a 9c 03 ba a5 eb 38 
	Digest:     32 8d 55 6c 94 85 a1 b6 0e 6b 76 6f 23 6e a6 d5 
	            08 e4 5d 7d 6b 02 93 7a 49 e0 9f 26 84 4c fb 17 

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Monday 20 January 2025  06:17:33 -0500 (0:00:00.433)       0:06:02.070 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Monday 20 January 2025  06:17:33 -0500 (0:00:00.046)       0:06:02.117 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Monday 20 January 2025  06:17:33 -0500 (0:00:00.043)       0:06:02.161 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Monday 20 January 2025  06:17:33 -0500 (0:00:00.040)       0:06:02.201 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Monday 20 January 2025  06:17:33 -0500 (0:00:00.048)       0:06:02.249 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Monday 20 January 2025  06:17:33 -0500 (0:00:00.051)       0:06:02.301 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Monday 20 January 2025  06:17:33 -0500 (0:00:00.034)       0:06:02.336 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Monday 20 January 2025  06:17:33 -0500 (0:00:00.032)       0:06:02.368 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [
            "luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81 /dev/sda1 VALUE_SPECIFIED_IN_NO_LOG_PARAMETER"
        ],
        "_storage_test_expected_crypttab_entries": "1",
        "_storage_test_expected_crypttab_key_file": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Monday 20 January 2025  06:17:33 -0500 (0:00:00.040)       0:06:02.408 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Monday 20 January 2025  06:17:33 -0500 (0:00:00.043)       0:06:02.452 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Monday 20 January 2025  06:17:33 -0500 (0:00:00.070)       0:06:02.523 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Monday 20 January 2025  06:17:33 -0500 (0:00:00.063)       0:06:02.586 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Monday 20 January 2025  06:17:33 -0500 (0:00:00.056)       0:06:02.643 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Monday 20 January 2025  06:17:33 -0500 (0:00:00.058)       0:06:02.701 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Monday 20 January 2025  06:17:33 -0500 (0:00:00.036)       0:06:02.738 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Monday 20 January 2025  06:17:34 -0500 (0:00:00.033)       0:06:02.771 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Monday 20 January 2025  06:17:34 -0500 (0:00:00.040)       0:06:02.812 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Monday 20 January 2025  06:17:34 -0500 (0:00:00.031)       0:06:02.844 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Monday 20 January 2025  06:17:34 -0500 (0:00:00.028)       0:06:02.872 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Monday 20 January 2025  06:17:34 -0500 (0:00:00.031)       0:06:02.904 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Monday 20 January 2025  06:17:34 -0500 (0:00:00.023)       0:06:02.928 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Monday 20 January 2025  06:17:34 -0500 (0:00:00.028)       0:06:02.956 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Monday 20 January 2025  06:17:34 -0500 (0:00:00.025)       0:06:02.981 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Monday 20 January 2025  06:17:34 -0500 (0:00:00.024)       0:06:03.006 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Monday 20 January 2025  06:17:34 -0500 (0:00:00.024)       0:06:03.030 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Monday 20 January 2025  06:17:34 -0500 (0:00:00.021)       0:06:03.052 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Monday 20 January 2025  06:17:34 -0500 (0:00:00.021)       0:06:03.074 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Monday 20 January 2025  06:17:34 -0500 (0:00:00.023)       0:06:03.097 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Monday 20 January 2025  06:17:34 -0500 (0:00:00.023)       0:06:03.121 ******** 
skipping: [managed-node1] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Monday 20 January 2025  06:17:34 -0500 (0:00:00.027)       0:06:03.149 ******** 
skipping: [managed-node1] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Monday 20 January 2025  06:17:34 -0500 (0:00:00.030)       0:06:03.179 ******** 
skipping: [managed-node1] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Monday 20 January 2025  06:17:34 -0500 (0:00:00.043)       0:06:03.222 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Monday 20 January 2025  06:17:34 -0500 (0:00:00.053)       0:06:03.275 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Monday 20 January 2025  06:17:34 -0500 (0:00:00.044)       0:06:03.320 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Monday 20 January 2025  06:17:34 -0500 (0:00:00.051)       0:06:03.371 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Monday 20 January 2025  06:17:34 -0500 (0:00:00.053)       0:06:03.425 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Monday 20 January 2025  06:17:34 -0500 (0:00:00.087)       0:06:03.513 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Monday 20 January 2025  06:17:34 -0500 (0:00:00.032)       0:06:03.545 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Monday 20 January 2025  06:17:34 -0500 (0:00:00.032)       0:06:03.578 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Monday 20 January 2025  06:17:34 -0500 (0:00:00.029)       0:06:03.607 ******** 
skipping: [managed-node1] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Monday 20 January 2025  06:17:34 -0500 (0:00:00.042)       0:06:03.650 ******** 
skipping: [managed-node1] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Monday 20 January 2025  06:17:34 -0500 (0:00:00.049)       0:06:03.699 ******** 
skipping: [managed-node1] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Monday 20 January 2025  06:17:34 -0500 (0:00:00.052)       0:06:03.752 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Monday 20 January 2025  06:17:35 -0500 (0:00:00.091)       0:06:03.844 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Monday 20 January 2025  06:17:35 -0500 (0:00:00.035)       0:06:03.879 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Monday 20 January 2025  06:17:35 -0500 (0:00:00.031)       0:06:03.911 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Monday 20 January 2025  06:17:35 -0500 (0:00:00.040)       0:06:03.952 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Monday 20 January 2025  06:17:35 -0500 (0:00:00.037)       0:06:03.989 ******** 
ok: [managed-node1] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Monday 20 January 2025  06:17:35 -0500 (0:00:00.045)       0:06:04.034 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Monday 20 January 2025  06:17:35 -0500 (0:00:00.047)       0:06:04.081 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Monday 20 January 2025  06:17:35 -0500 (0:00:00.035)       0:06:04.117 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Monday 20 January 2025  06:17:35 -0500 (0:00:00.034)       0:06:04.152 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Monday 20 January 2025  06:17:35 -0500 (0:00:00.041)       0:06:04.193 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Monday 20 January 2025  06:17:35 -0500 (0:00:00.032)       0:06:04.225 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Monday 20 January 2025  06:17:35 -0500 (0:00:00.025)       0:06:04.251 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Monday 20 January 2025  06:17:35 -0500 (0:00:00.031)       0:06:04.283 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Monday 20 January 2025  06:17:35 -0500 (0:00:00.024)       0:06:04.308 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Monday 20 January 2025  06:17:35 -0500 (0:00:00.027)       0:06:04.335 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Monday 20 January 2025  06:17:35 -0500 (0:00:00.024)       0:06:04.359 ******** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Monday 20 January 2025  06:17:35 -0500 (0:00:00.022)       0:06:04.381 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Remove the key file] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:358
Monday 20 January 2025  06:17:35 -0500 (0:00:00.027)       0:06:04.409 ******** 
ok: [managed-node1] => {
    "changed": false,
    "path": "/tmp/storage_test1_y4m_lvlukskey",
    "state": "absent"
}

TASK [Test for correct handling of new encrypted volume w/ no key] *************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:368
Monday 20 January 2025  06:17:36 -0500 (0:00:00.368)       0:06:04.778 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node1

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Monday 20 January 2025  06:17:36 -0500 (0:00:00.037)       0:06:04.816 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_pools_global": [],
        "storage_safe_mode_global": true,
        "storage_volumes_global": []
    },
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Monday 20 January 2025  06:17:36 -0500 (0:00:00.027)       0:06:04.843 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:17:36 -0500 (0:00:00.029)       0:06:04.872 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:17:36 -0500 (0:00:00.030)       0:06:04.903 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:17:36 -0500 (0:00:00.023)       0:06:04.927 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:17:36 -0500 (0:00:00.055)       0:06:04.982 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:17:36 -0500 (0:00:00.020)       0:06:05.003 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:17:36 -0500 (0:00:00.019)       0:06:05.023 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:17:36 -0500 (0:00:00.018)       0:06:05.041 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:17:36 -0500 (0:00:00.019)       0:06:05.061 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:17:36 -0500 (0:00:00.049)       0:06:05.110 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:17:39 -0500 (0:00:02.893)       0:06:08.004 ******** 
ok: [managed-node1] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "encryption": true,
                    "encryption_luks_version": "luks2",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:17:39 -0500 (0:00:00.052)       0:06:08.056 ******** 
ok: [managed-node1] => {
    "storage_volumes": []
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:17:39 -0500 (0:00:00.066)       0:06:08.123 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "cryptsetup",
        "lvm2"
    ],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:17:43 -0500 (0:00:04.443)       0:06:12.566 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:17:43 -0500 (0:00:00.114)       0:06:12.680 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:17:44 -0500 (0:00:00.091)       0:06:12.772 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:17:44 -0500 (0:00:00.056)       0:06:12.828 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:17:44 -0500 (0:00:00.062)       0:06:12.890 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:17:47 -0500 (0:00:03.044)       0:06:15.935 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:17:48 -0500 (0:00:01.662)       0:06:17.598 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:17:48 -0500 (0:00:00.034)       0:06:17.632 ******** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:17:48 -0500 (0:00:00.019)       0:06:17.651 ******** 
fatal: [managed-node1]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

encrypted volume 'test1' missing key/password

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Monday 20 January 2025  06:17:53 -0500 (0:00:04.241)       0:06:21.893 ******** 
fatal: [managed-node1]: FAILED! => {
    "changed": false
}

MSG:

{'msg': "encrypted volume 'test1' missing key/password", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'encryption_clevis_pin': None, 'encryption_tang_url': None, 'encryption_tang_thumbprint': None, 'grow_to_fill': False, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'shared': False, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': True, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': 'luks2', 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}]}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'grow_to_fill': False, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, 'shared': False}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': False, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:17:53 -0500 (0:00:00.043)       0:06:21.936 ******** 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Monday 20 January 2025  06:17:53 -0500 (0:00:00.034)       0:06:21.971 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Monday 20 January 2025  06:17:53 -0500 (0:00:00.040)       0:06:22.011 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Monday 20 January 2025  06:17:53 -0500 (0:00:00.045)       0:06:22.057 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Create an encrypted lvm volume w/ default fs] ****************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:387
Monday 20 January 2025  06:17:53 -0500 (0:00:00.036)       0:06:22.093 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:17:53 -0500 (0:00:00.081)       0:06:22.175 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:17:53 -0500 (0:00:00.100)       0:06:22.275 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:17:53 -0500 (0:00:00.045)       0:06:22.321 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:17:53 -0500 (0:00:00.148)       0:06:22.469 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:17:53 -0500 (0:00:00.040)       0:06:22.510 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:17:53 -0500 (0:00:00.045)       0:06:22.556 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:17:53 -0500 (0:00:00.038)       0:06:22.594 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:17:53 -0500 (0:00:00.071)       0:06:22.666 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:17:54 -0500 (0:00:00.120)       0:06:22.786 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:17:57 -0500 (0:00:03.197)       0:06:25.984 ******** 
ok: [managed-node1] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "encryption": true,
                    "encryption_cipher": "aes-xts-plain64",
                    "encryption_key_size": 512,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "yabbadabbadoo",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:17:57 -0500 (0:00:00.030)       0:06:26.015 ******** 
ok: [managed-node1] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:17:57 -0500 (0:00:00.021)       0:06:26.036 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "cryptsetup",
        "lvm2"
    ],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:18:01 -0500 (0:00:04.201)       0:06:30.238 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:18:01 -0500 (0:00:00.054)       0:06:30.293 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:18:01 -0500 (0:00:00.030)       0:06:30.323 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:18:01 -0500 (0:00:00.031)       0:06:30.355 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:18:01 -0500 (0:00:00.031)       0:06:30.386 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:18:04 -0500 (0:00:02.899)       0:06:33.285 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:18:06 -0500 (0:00:01.706)       0:06:34.992 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:18:06 -0500 (0:00:00.041)       0:06:35.033 ******** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:18:06 -0500 (0:00:00.079)       0:06:35.113 ******** 
changed: [managed-node1] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
            "fs_type": "xfs"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/sda1",
            "fs_type": "luks"
        },
        {
            "action": "destroy device",
            "device": "/dev/sda1",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/sda",
            "fs_type": "disklabel"
        },
        {
            "action": "create format",
            "device": "/dev/sda",
            "fs_type": "lvmpv"
        },
        {
            "action": "create device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "create device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "luks"
        },
        {
            "action": "create device",
            "device": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
            "fs_type": "xfs"
        }
    ],
    "changed": true,
    "crypts": [
        {
            "backing_device": "/dev/sda1",
            "name": "luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
            "password": "-",
            "state": "absent"
        },
        {
            "backing_device": "/dev/mapper/foo-test1",
            "name": "luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
            "password": "-",
            "state": "present"
        }
    ],
    "leaves": [
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1",
        "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6"
    ],
    "mounts": [
        {
            "fstype": "xfs",
            "path": "/opt/test1",
            "src": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
            "state": "absent"
        },
        {
            "dump": 0,
            "fstype": "xfs",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "lvm2",
        "cryptsetup"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                    "_kernel_device": "/dev/dm-1",
                    "_mount_id": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": true,
                    "encryption_cipher": "aes-xts-plain64",
                    "encryption_key": null,
                    "encryption_key_size": 512,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "xfs",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Monday 20 January 2025  06:18:21 -0500 (0:00:15.262)       0:06:50.376 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Monday 20 January 2025  06:18:21 -0500 (0:00:00.031)       0:06:50.407 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371840.9730477,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "ecd570e52342bc865a4fbda5ac5097f233ee8fb1",
        "ctime": 1737371840.9700477,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 207618244,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1737371840.9700477,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1425,
        "uid": 0,
        "version": "261612921",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Monday 20 January 2025  06:18:22 -0500 (0:00:00.401)       0:06:50.808 ******** 
ok: [managed-node1] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:18:22 -0500 (0:00:00.415)       0:06:51.224 ******** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Monday 20 January 2025  06:18:22 -0500 (0:00:00.028)       0:06:51.252 ******** 
ok: [managed-node1] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
                "fs_type": "xfs"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/sda1",
                "fs_type": "luks"
            },
            {
                "action": "destroy device",
                "device": "/dev/sda1",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/sda",
                "fs_type": "disklabel"
            },
            {
                "action": "create format",
                "device": "/dev/sda",
                "fs_type": "lvmpv"
            },
            {
                "action": "create device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "create device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "luks"
            },
            {
                "action": "create device",
                "device": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                "fs_type": "xfs"
            }
        ],
        "changed": true,
        "crypts": [
            {
                "backing_device": "/dev/sda1",
                "name": "luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
                "password": "-",
                "state": "absent"
            },
            {
                "backing_device": "/dev/mapper/foo-test1",
                "name": "luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                "password": "-",
                "state": "present"
            }
        ],
        "failed": false,
        "leaves": [
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1",
            "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6"
        ],
        "mounts": [
            {
                "fstype": "xfs",
                "path": "/opt/test1",
                "src": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
                "state": "absent"
            },
            {
                "dump": 0,
                "fstype": "xfs",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "lvm2",
            "cryptsetup"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                        "_kernel_device": "/dev/dm-1",
                        "_mount_id": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": true,
                        "encryption_cipher": "aes-xts-plain64",
                        "encryption_key": null,
                        "encryption_key_size": 512,
                        "encryption_luks_version": "luks2",
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "xfs",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Monday 20 January 2025  06:18:22 -0500 (0:00:00.041)       0:06:51.294 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                        "_kernel_device": "/dev/dm-1",
                        "_mount_id": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": true,
                        "encryption_cipher": "aes-xts-plain64",
                        "encryption_key": null,
                        "encryption_key_size": 512,
                        "encryption_luks_version": "luks2",
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "xfs",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Monday 20 January 2025  06:18:22 -0500 (0:00:00.046)       0:06:51.341 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Monday 20 January 2025  06:18:22 -0500 (0:00:00.036)       0:06:51.378 ******** 
changed: [managed-node1] => (item={'src': '/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "fstype": "xfs",
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Monday 20 January 2025  06:18:23 -0500 (0:00:00.422)       0:06:51.800 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Monday 20 January 2025  06:18:23 -0500 (0:00:00.688)       0:06:52.489 ******** 
changed: [managed-node1] => (item={'src': '/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Monday 20 January 2025  06:18:24 -0500 (0:00:00.445)       0:06:52.934 ******** 
skipping: [managed-node1] => (item={'src': '/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Monday 20 January 2025  06:18:24 -0500 (0:00:00.042)       0:06:52.977 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Monday 20 January 2025  06:18:24 -0500 (0:00:00.739)       0:06:53.717 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371845.484056,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "a3cc9ce6251dfc5492cb566f723e0633bb00243b",
        "ctime": 1737371842.8500512,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 325058693,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0600",
        "mtime": 1737371842.8490512,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 88,
        "uid": 0,
        "version": "621844829",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Monday 20 January 2025  06:18:25 -0500 (0:00:00.535)       0:06:54.252 ******** 
changed: [managed-node1] => (item={'backing_device': '/dev/sda1', 'name': 'luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81', 'password': '-', 'state': 'absent'}) => {
    "ansible_loop_var": "entry",
    "backup": "",
    "changed": true,
    "entry": {
        "backing_device": "/dev/sda1",
        "name": "luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
        "password": "-",
        "state": "absent"
    },
    "found": 1
}

MSG:

1 line(s) removed
changed: [managed-node1] => (item={'backing_device': '/dev/mapper/foo-test1', 'name': 'luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6', 'password': '-', 'state': 'present'}) => {
    "ansible_loop_var": "entry",
    "backup": "",
    "changed": true,
    "entry": {
        "backing_device": "/dev/mapper/foo-test1",
        "name": "luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
        "password": "-",
        "state": "present"
    }
}

MSG:

line added

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Monday 20 January 2025  06:18:26 -0500 (0:00:00.858)       0:06:55.110 ******** 
ok: [managed-node1]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:406
Monday 20 January 2025  06:18:27 -0500 (0:00:00.751)       0:06:55.862 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node1

TASK [Print out pool information] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Monday 20 January 2025  06:18:27 -0500 (0:00:00.035)       0:06:55.898 ******** 
ok: [managed-node1] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                    "_kernel_device": "/dev/dm-1",
                    "_mount_id": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": true,
                    "encryption_cipher": "aes-xts-plain64",
                    "encryption_key": null,
                    "encryption_key_size": 512,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "xfs",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Monday 20 January 2025  06:18:27 -0500 (0:00:00.027)       0:06:55.925 ******** 
skipping: [managed-node1] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Monday 20 January 2025  06:18:27 -0500 (0:00:00.019)       0:06:55.945 ******** 
ok: [managed-node1] => {
    "changed": false,
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "crypto_LUKS",
            "label": "",
            "mountpoint": "",
            "name": "/dev/mapper/foo-test1",
            "size": "4G",
            "type": "lvm",
            "uuid": "940b35f5-2180-4a53-8562-e5b2ccb3efe6"
        },
        "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
            "size": "4G",
            "type": "crypt",
            "uuid": "9d7054c7-e815-4da1-9cee-d4e79336c702"
        },
        "/dev/sda": {
            "fstype": "LVM2_member",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "kSV36d-iWVA-92sg-L6wf-6hE7-Eik8-GLOUAJ"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Monday 20 January 2025  06:18:27 -0500 (0:00:00.356)       0:06:56.301 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.002551",
    "end": "2025-01-20 06:18:27.830022",
    "rc": 0,
    "start": "2025-01-20 06:18:27.827471"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6 /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Monday 20 January 2025  06:18:27 -0500 (0:00:00.349)       0:06:56.651 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003073",
    "end": "2025-01-20 06:18:28.180894",
    "failed_when_result": false,
    "rc": 0,
    "start": "2025-01-20 06:18:28.177821"
}

STDOUT:

luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6 /dev/mapper/foo-test1 -

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Monday 20 January 2025  06:18:28 -0500 (0:00:00.350)       0:06:57.002 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node1

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Monday 20 January 2025  06:18:28 -0500 (0:00:00.041)       0:06:57.044 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Monday 20 January 2025  06:18:28 -0500 (0:00:00.056)       0:06:57.100 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "vgs",
        "--noheadings",
        "--binary",
        "-o",
        "shared",
        "foo"
    ],
    "delta": "0:00:00.023968",
    "end": "2025-01-20 06:18:28.659777",
    "rc": 0,
    "start": "2025-01-20 06:18:28.635809"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Monday 20 January 2025  06:18:28 -0500 (0:00:00.388)       0:06:57.489 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Monday 20 January 2025  06:18:28 -0500 (0:00:00.030)       0:06:57.519 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node1

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Monday 20 January 2025  06:18:28 -0500 (0:00:00.050)       0:06:57.570 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Monday 20 January 2025  06:18:28 -0500 (0:00:00.043)       0:06:57.613 ******** 
ok: [managed-node1] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sda",
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Monday 20 January 2025  06:18:29 -0500 (0:00:00.520)       0:06:58.133 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Monday 20 January 2025  06:18:29 -0500 (0:00:00.038)       0:06:58.171 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Monday 20 January 2025  06:18:29 -0500 (0:00:00.043)       0:06:58.215 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Monday 20 January 2025  06:18:29 -0500 (0:00:00.037)       0:06:58.253 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Monday 20 January 2025  06:18:29 -0500 (0:00:00.038)       0:06:58.291 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Monday 20 January 2025  06:18:29 -0500 (0:00:00.038)       0:06:58.330 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Monday 20 January 2025  06:18:29 -0500 (0:00:00.035)       0:06:58.365 ******** 
ok: [managed-node1] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Monday 20 January 2025  06:18:29 -0500 (0:00:00.046)       0:06:58.412 ******** 
ok: [managed-node1] => {
    "changed": false,
    "failed_when_result": false,
    "rc": 1
}

STDERR:

Shared connection to 10.31.8.138 closed.



MSG:

non-zero return code

TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:74
Monday 20 January 2025  06:18:30 -0500 (0:00:00.471)       0:06:58.884 ******** 
skipping: [managed-node1] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:84
Monday 20 January 2025  06:18:30 -0500 (0:00:00.043)       0:06:58.928 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node1

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Monday 20 January 2025  06:18:30 -0500 (0:00:00.074)       0:06:59.002 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Monday 20 January 2025  06:18:30 -0500 (0:00:00.038)       0:06:59.041 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Monday 20 January 2025  06:18:30 -0500 (0:00:00.031)       0:06:59.072 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Monday 20 January 2025  06:18:30 -0500 (0:00:00.031)       0:06:59.103 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Monday 20 January 2025  06:18:30 -0500 (0:00:00.034)       0:06:59.138 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Monday 20 January 2025  06:18:30 -0500 (0:00:00.030)       0:06:59.168 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Monday 20 January 2025  06:18:30 -0500 (0:00:00.032)       0:06:59.201 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Monday 20 January 2025  06:18:30 -0500 (0:00:00.025)       0:06:59.226 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Monday 20 January 2025  06:18:30 -0500 (0:00:00.025)       0:06:59.251 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Monday 20 January 2025  06:18:30 -0500 (0:00:00.023)       0:06:59.275 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Monday 20 January 2025  06:18:30 -0500 (0:00:00.023)       0:06:59.298 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:87
Monday 20 January 2025  06:18:30 -0500 (0:00:00.020)       0:06:59.319 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node1

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Monday 20 January 2025  06:18:30 -0500 (0:00:00.063)       0:06:59.382 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node1

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Monday 20 January 2025  06:18:30 -0500 (0:00:00.066)       0:06:59.449 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Monday 20 January 2025  06:18:30 -0500 (0:00:00.033)       0:06:59.483 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Monday 20 January 2025  06:18:30 -0500 (0:00:00.036)       0:06:59.519 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Monday 20 January 2025  06:18:30 -0500 (0:00:00.036)       0:06:59.555 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Monday 20 January 2025  06:18:30 -0500 (0:00:00.033)       0:06:59.589 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Monday 20 January 2025  06:18:30 -0500 (0:00:00.026)       0:06:59.615 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Monday 20 January 2025  06:18:30 -0500 (0:00:00.045)       0:06:59.661 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:90
Monday 20 January 2025  06:18:30 -0500 (0:00:00.041)       0:06:59.702 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node1

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Monday 20 January 2025  06:18:30 -0500 (0:00:00.058)       0:06:59.761 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node1

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Monday 20 January 2025  06:18:31 -0500 (0:00:00.050)       0:06:59.811 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Monday 20 January 2025  06:18:31 -0500 (0:00:00.025)       0:06:59.837 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Monday 20 January 2025  06:18:31 -0500 (0:00:00.022)       0:06:59.860 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Monday 20 January 2025  06:18:31 -0500 (0:00:00.063)       0:06:59.923 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:93
Monday 20 January 2025  06:18:31 -0500 (0:00:00.022)       0:06:59.945 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node1

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Monday 20 January 2025  06:18:31 -0500 (0:00:00.045)       0:06:59.990 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Monday 20 January 2025  06:18:31 -0500 (0:00:00.025)       0:07:00.016 ******** 
skipping: [managed-node1] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Monday 20 January 2025  06:18:31 -0500 (0:00:00.032)       0:07:00.049 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node1

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Monday 20 January 2025  06:18:31 -0500 (0:00:00.046)       0:07:00.095 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Monday 20 January 2025  06:18:31 -0500 (0:00:00.064)       0:07:00.160 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Monday 20 January 2025  06:18:31 -0500 (0:00:00.062)       0:07:00.222 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Monday 20 January 2025  06:18:31 -0500 (0:00:00.038)       0:07:00.261 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Monday 20 January 2025  06:18:31 -0500 (0:00:00.032)       0:07:00.293 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Monday 20 January 2025  06:18:31 -0500 (0:00:00.031)       0:07:00.325 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Monday 20 January 2025  06:18:31 -0500 (0:00:00.030)       0:07:00.355 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:96
Monday 20 January 2025  06:18:31 -0500 (0:00:00.027)       0:07:00.382 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node1

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Monday 20 January 2025  06:18:31 -0500 (0:00:00.058)       0:07:00.441 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node1

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Monday 20 January 2025  06:18:31 -0500 (0:00:00.052)       0:07:00.493 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Monday 20 January 2025  06:18:31 -0500 (0:00:00.023)       0:07:00.517 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Monday 20 January 2025  06:18:31 -0500 (0:00:00.025)       0:07:00.543 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Monday 20 January 2025  06:18:31 -0500 (0:00:00.020)       0:07:00.564 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Monday 20 January 2025  06:18:31 -0500 (0:00:00.023)       0:07:00.588 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Monday 20 January 2025  06:18:31 -0500 (0:00:00.035)       0:07:00.623 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Monday 20 January 2025  06:18:31 -0500 (0:00:00.024)       0:07:00.648 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:99
Monday 20 January 2025  06:18:31 -0500 (0:00:00.039)       0:07:00.687 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node1

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Monday 20 January 2025  06:18:32 -0500 (0:00:00.079)       0:07:00.767 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Monday 20 January 2025  06:18:32 -0500 (0:00:00.048)       0:07:00.815 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Monday 20 January 2025  06:18:32 -0500 (0:00:00.053)       0:07:00.869 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Monday 20 January 2025  06:18:32 -0500 (0:00:00.040)       0:07:00.910 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Monday 20 January 2025  06:18:32 -0500 (0:00:00.050)       0:07:00.961 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Monday 20 January 2025  06:18:32 -0500 (0:00:00.053)       0:07:01.015 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:102
Monday 20 January 2025  06:18:32 -0500 (0:00:00.072)       0:07:01.088 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Monday 20 January 2025  06:18:32 -0500 (0:00:00.065)       0:07:01.154 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node1

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Monday 20 January 2025  06:18:32 -0500 (0:00:00.073)       0:07:01.227 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Monday 20 January 2025  06:18:32 -0500 (0:00:00.051)       0:07:01.278 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node1

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Monday 20 January 2025  06:18:32 -0500 (0:00:00.172)       0:07:01.450 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Monday 20 January 2025  06:18:32 -0500 (0:00:00.038)       0:07:01.489 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Monday 20 January 2025  06:18:32 -0500 (0:00:00.039)       0:07:01.529 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Monday 20 January 2025  06:18:32 -0500 (0:00:00.032)       0:07:01.561 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Monday 20 January 2025  06:18:32 -0500 (0:00:00.037)       0:07:01.599 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Monday 20 January 2025  06:18:32 -0500 (0:00:00.031)       0:07:01.630 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Monday 20 January 2025  06:18:32 -0500 (0:00:00.056)       0:07:01.686 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Monday 20 January 2025  06:18:32 -0500 (0:00:00.046)       0:07:01.733 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Monday 20 January 2025  06:18:33 -0500 (0:00:00.048)       0:07:01.782 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Monday 20 January 2025  06:18:33 -0500 (0:00:00.040)       0:07:01.822 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Monday 20 January 2025  06:18:33 -0500 (0:00:00.037)       0:07:01.860 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Monday 20 January 2025  06:18:33 -0500 (0:00:00.031)       0:07:01.891 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Monday 20 January 2025  06:18:33 -0500 (0:00:00.062)       0:07:01.954 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Monday 20 January 2025  06:18:33 -0500 (0:00:00.040)       0:07:01.995 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Monday 20 January 2025  06:18:33 -0500 (0:00:00.043)       0:07:02.038 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Monday 20 January 2025  06:18:33 -0500 (0:00:00.114)       0:07:02.152 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Monday 20 January 2025  06:18:33 -0500 (0:00:00.040)       0:07:02.193 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Monday 20 January 2025  06:18:33 -0500 (0:00:00.030)       0:07:02.223 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Monday 20 January 2025  06:18:33 -0500 (0:00:00.042)       0:07:02.266 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Monday 20 January 2025  06:18:33 -0500 (0:00:00.045)       0:07:02.312 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371901.3091884,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1737371901.3091884,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 671645,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1737371901.3091884,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Monday 20 January 2025  06:18:33 -0500 (0:00:00.426)       0:07:02.738 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Monday 20 January 2025  06:18:34 -0500 (0:00:00.042)       0:07:02.781 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Monday 20 January 2025  06:18:34 -0500 (0:00:00.036)       0:07:02.818 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Monday 20 January 2025  06:18:34 -0500 (0:00:00.063)       0:07:02.881 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Monday 20 January 2025  06:18:34 -0500 (0:00:00.044)       0:07:02.926 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Monday 20 January 2025  06:18:34 -0500 (0:00:00.048)       0:07:02.975 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Monday 20 January 2025  06:18:34 -0500 (0:00:00.039)       0:07:03.015 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371901.4491887,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1737371901.4491887,
        "dev": 6,
        "device_type": 64769,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 672782,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1737371901.4491887,
        "nlink": 1,
        "path": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Monday 20 January 2025  06:18:34 -0500 (0:00:00.455)       0:07:03.470 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Monday 20 January 2025  06:18:37 -0500 (0:00:02.943)       0:07:06.414 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cryptsetup",
        "luksDump",
        "/dev/mapper/foo-test1"
    ],
    "delta": "0:00:00.009870",
    "end": "2025-01-20 06:18:38.001726",
    "rc": 0,
    "start": "2025-01-20 06:18:37.991856"
}

STDOUT:

LUKS header information
Version:       	2
Epoch:         	3
Metadata area: 	16384 [bytes]
Keyslots area: 	16744448 [bytes]
UUID:          	940b35f5-2180-4a53-8562-e5b2ccb3efe6
Label:         	(no label)
Subsystem:     	(no subsystem)
Flags:       	(no flags)

Data segments:
  0: crypt
	offset: 16777216 [bytes]
	length: (whole device)
	cipher: aes-xts-plain64
	sector: 512 [bytes]

Keyslots:
  0: luks2
	Key:        512 bits
	Priority:   normal
	Cipher:     aes-xts-plain64
	Cipher key: 512 bits
	PBKDF:      argon2i
	Time cost:  4
	Memory:     786697
	Threads:    2
	Salt:       91 2f 2a 22 b6 0e 7a bd 9f 87 7d 5d 23 fe ac 1b 
	            ad 4f fd 0b 88 3c ac d7 3e 11 7a 92 95 fd 43 8d 
	AF stripes: 4000
	AF hash:    sha256
	Area offset:32768 [bytes]
	Area length:258048 [bytes]
	Digest ID:  0
Tokens:
Digests:
  0: pbkdf2
	Hash:       sha256
	Iterations: 120249
	Salt:       7d 7b e3 a9 e1 08 76 be 61 e0 d0 f1 8e fd f4 89 
	            ed 05 7e c2 0d a2 41 7a 7a 49 f0 17 3b 68 5c 74 
	Digest:     7d 70 07 42 f6 fe 4f 52 91 9e 37 e3 42 46 62 2a 
	            fa ec e0 d3 fa 94 bf 70 90 4e de db f9 af e1 c7 

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Monday 20 January 2025  06:18:38 -0500 (0:00:00.444)       0:07:06.859 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Monday 20 January 2025  06:18:38 -0500 (0:00:00.038)       0:07:06.897 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Monday 20 January 2025  06:18:38 -0500 (0:00:00.040)       0:07:06.937 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Monday 20 January 2025  06:18:38 -0500 (0:00:00.044)       0:07:06.982 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Monday 20 January 2025  06:18:38 -0500 (0:00:00.042)       0:07:07.024 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Monday 20 January 2025  06:18:38 -0500 (0:00:00.038)       0:07:07.062 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Monday 20 January 2025  06:18:38 -0500 (0:00:00.033)       0:07:07.096 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Monday 20 January 2025  06:18:38 -0500 (0:00:00.037)       0:07:07.133 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [
            "luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6 /dev/mapper/foo-test1 -"
        ],
        "_storage_test_expected_crypttab_entries": "1",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Monday 20 January 2025  06:18:38 -0500 (0:00:00.031)       0:07:07.164 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Monday 20 January 2025  06:18:38 -0500 (0:00:00.030)       0:07:07.195 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Monday 20 January 2025  06:18:38 -0500 (0:00:00.030)       0:07:07.225 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Monday 20 January 2025  06:18:38 -0500 (0:00:00.027)       0:07:07.253 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Monday 20 January 2025  06:18:38 -0500 (0:00:00.027)       0:07:07.280 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Monday 20 January 2025  06:18:38 -0500 (0:00:00.020)       0:07:07.301 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Monday 20 January 2025  06:18:38 -0500 (0:00:00.019)       0:07:07.321 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Monday 20 January 2025  06:18:38 -0500 (0:00:00.023)       0:07:07.344 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Monday 20 January 2025  06:18:38 -0500 (0:00:00.019)       0:07:07.363 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Monday 20 January 2025  06:18:38 -0500 (0:00:00.020)       0:07:07.384 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Monday 20 January 2025  06:18:38 -0500 (0:00:00.019)       0:07:07.404 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Monday 20 January 2025  06:18:38 -0500 (0:00:00.020)       0:07:07.424 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Monday 20 January 2025  06:18:38 -0500 (0:00:00.018)       0:07:07.443 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Monday 20 January 2025  06:18:38 -0500 (0:00:00.022)       0:07:07.465 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Monday 20 January 2025  06:18:38 -0500 (0:00:00.020)       0:07:07.485 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Monday 20 January 2025  06:18:38 -0500 (0:00:00.020)       0:07:07.505 ******** 
ok: [managed-node1] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Monday 20 January 2025  06:18:39 -0500 (0:00:00.481)       0:07:07.987 ******** 
ok: [managed-node1] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Monday 20 January 2025  06:18:39 -0500 (0:00:00.408)       0:07:08.396 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_expected_size": "4294967296"
    },
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Monday 20 January 2025  06:18:39 -0500 (0:00:00.044)       0:07:08.440 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Monday 20 January 2025  06:18:39 -0500 (0:00:00.033)       0:07:08.473 ******** 
ok: [managed-node1] => {
    "bytes": 10726680821,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Monday 20 January 2025  06:18:40 -0500 (0:00:00.415)       0:07:08.889 ******** 
skipping: [managed-node1] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Monday 20 January 2025  06:18:40 -0500 (0:00:00.044)       0:07:08.933 ******** 
skipping: [managed-node1] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Monday 20 January 2025  06:18:40 -0500 (0:00:00.036)       0:07:08.970 ******** 
skipping: [managed-node1] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Monday 20 January 2025  06:18:40 -0500 (0:00:00.036)       0:07:09.006 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Monday 20 January 2025  06:18:40 -0500 (0:00:00.035)       0:07:09.042 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Monday 20 January 2025  06:18:40 -0500 (0:00:00.031)       0:07:09.074 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Monday 20 January 2025  06:18:40 -0500 (0:00:00.029)       0:07:09.104 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Monday 20 January 2025  06:18:40 -0500 (0:00:00.030)       0:07:09.134 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Monday 20 January 2025  06:18:40 -0500 (0:00:00.031)       0:07:09.165 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Monday 20 January 2025  06:18:40 -0500 (0:00:00.029)       0:07:09.195 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Monday 20 January 2025  06:18:40 -0500 (0:00:00.030)       0:07:09.226 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Monday 20 January 2025  06:18:40 -0500 (0:00:00.030)       0:07:09.256 ******** 
skipping: [managed-node1] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Monday 20 January 2025  06:18:40 -0500 (0:00:00.031)       0:07:09.288 ******** 
skipping: [managed-node1] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Monday 20 January 2025  06:18:40 -0500 (0:00:00.031)       0:07:09.319 ******** 
skipping: [managed-node1] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Monday 20 January 2025  06:18:40 -0500 (0:00:00.032)       0:07:09.351 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Monday 20 January 2025  06:18:40 -0500 (0:00:00.032)       0:07:09.384 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Monday 20 January 2025  06:18:40 -0500 (0:00:00.024)       0:07:09.408 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Monday 20 January 2025  06:18:40 -0500 (0:00:00.024)       0:07:09.433 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Monday 20 January 2025  06:18:40 -0500 (0:00:00.021)       0:07:09.454 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Monday 20 January 2025  06:18:40 -0500 (0:00:00.027)       0:07:09.482 ******** 
ok: [managed-node1] => {
    "storage_test_actual_size": {
        "bytes": 4294967296,
        "changed": false,
        "failed": false,
        "lvm": "4g",
        "parted": "4GiB",
        "size": "4 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Monday 20 January 2025  06:18:40 -0500 (0:00:00.068)       0:07:09.550 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Monday 20 January 2025  06:18:40 -0500 (0:00:00.032)       0:07:09.583 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Monday 20 January 2025  06:18:40 -0500 (0:00:00.033)       0:07:09.616 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.023999",
    "end": "2025-01-20 06:18:41.187305",
    "rc": 0,
    "start": "2025-01-20 06:18:41.163306"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Monday 20 January 2025  06:18:41 -0500 (0:00:00.396)       0:07:10.012 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Monday 20 January 2025  06:18:41 -0500 (0:00:00.027)       0:07:10.040 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Monday 20 January 2025  06:18:41 -0500 (0:00:00.029)       0:07:10.070 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Monday 20 January 2025  06:18:41 -0500 (0:00:00.024)       0:07:10.094 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Monday 20 January 2025  06:18:41 -0500 (0:00:00.026)       0:07:10.120 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Monday 20 January 2025  06:18:41 -0500 (0:00:00.024)       0:07:10.144 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Monday 20 January 2025  06:18:41 -0500 (0:00:00.027)       0:07:10.172 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Monday 20 January 2025  06:18:41 -0500 (0:00:00.025)       0:07:10.197 ******** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Monday 20 January 2025  06:18:41 -0500 (0:00:00.028)       0:07:10.225 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Verify preservation of encryption settings on existing LVM volume] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:409
Monday 20 January 2025  06:18:41 -0500 (0:00:00.047)       0:07:10.273 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:18:41 -0500 (0:00:00.082)       0:07:10.355 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:18:41 -0500 (0:00:00.048)       0:07:10.403 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:18:41 -0500 (0:00:00.031)       0:07:10.434 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:18:41 -0500 (0:00:00.067)       0:07:10.501 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:18:41 -0500 (0:00:00.027)       0:07:10.529 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:18:41 -0500 (0:00:00.025)       0:07:10.555 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:18:41 -0500 (0:00:00.024)       0:07:10.579 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:18:41 -0500 (0:00:00.025)       0:07:10.605 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:18:41 -0500 (0:00:00.070)       0:07:10.676 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:18:44 -0500 (0:00:02.938)       0:07:13.614 ******** 
ok: [managed-node1] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:18:44 -0500 (0:00:00.046)       0:07:13.661 ******** 
ok: [managed-node1] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:18:44 -0500 (0:00:00.035)       0:07:13.697 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "lvm2"
    ],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:18:49 -0500 (0:00:04.227)       0:07:17.925 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:18:49 -0500 (0:00:00.058)       0:07:17.984 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:18:49 -0500 (0:00:00.039)       0:07:18.023 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:18:49 -0500 (0:00:00.024)       0:07:18.048 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:18:49 -0500 (0:00:00.025)       0:07:18.074 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:18:52 -0500 (0:00:02.922)       0:07:20.996 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-cryptsetup@luk...dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service": {
                "name": "systemd-cryptsetup@luk...dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
                "source": "systemd",
                "state": "inactive",
                "status": "generated"
            },
            "systemd-cryptsetup@luks\\x2d4fd96f37\\x2dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service": {
                "name": "systemd-cryptsetup@luks\\x2d4fd96f37\\x2dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:18:54 -0500 (0:00:01.792)       0:07:22.789 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": [
            "systemd-cryptsetup@luks\\x2d4fd96f37\\x2dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
            "systemd-cryptsetup@luk...dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service"
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:18:54 -0500 (0:00:00.041)       0:07:22.831 ******** 
changed: [managed-node1] => (item=systemd-cryptsetup@luks\x2d4fd96f37\x2dc9c1\x2d41aa\x2dbce3\x2d48eefe3c9f81.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luks\\x2d4fd96f37\\x2dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
    "name": "systemd-cryptsetup@luks\\x2d4fd96f37\\x2dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "After": "dev-sda1.device system-systemd\\x2dcryptsetup.slice systemd-journald.socket cryptsetup-pre.target",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "Before": "cryptsetup.target umount.target",
        "BindsTo": "dev-sda1.device",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "yes",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "Conflicts": "umount.target",
        "ControlPID": "0",
        "DefaultDependencies": "no",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "Cryptography Setup for luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81",
        "DevicePolicy": "auto",
        "Documentation": "man:crypttab(5) man:systemd-cryptsetup-generator(8) man:systemd-cryptsetup@.service(8)",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81 /dev/sda1 VALUE_SPECIFIED_IN_NO_LOG_PARAMETER  ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
        "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-4fd96f37-c9c1-41aa-bce3-48eefe3c9f81 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2d4fd96f37\\x2dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luks\\x2d4fd96f37\\x2dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
        "IgnoreOnIsolate": "yes",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "shared",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "0",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "65536",
        "LimitMEMLOCKSoft": "65536",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "262144",
        "LimitNOFILESoft": "1024",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadState": "loaded",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luks\\x2d4fd96f37\\x2dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
        "NeedDaemonReload": "yes",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "yes",
        "RemoveIPC": "no",
        "RequiredBy": "cryptsetup.target",
        "Requires": "system-systemd\\x2dcryptsetup.slice",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "Slice": "system-systemd\\x2dcryptsetup.slice",
        "SourcePath": "/etc/crypttab",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "journal",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestamp": "Mon 2025-01-20 06:18:24 EST",
        "StateChangeTimestampMonotonic": "2722718376",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "infinity",
        "TimeoutStopUSec": "infinity",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "Type": "oneshot",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFilePreset": "disabled",
        "UnitFileState": "generated",
        "UtmpMode": "init",
        "WantedBy": "dev-sda1.device",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}
changed: [managed-node1] => (item=systemd-cryptsetup@luk...dc9c1\x2d41aa\x2dbce3\x2d48eefe3c9f81.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luk...dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
    "name": "systemd-cryptsetup@luk...dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luk...dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luk...dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.NoSuchUnit \"Unit systemd-cryptsetup@luk...dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service not found.\"",
        "LoadState": "not-found",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luk...dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:18:55 -0500 (0:00:01.354)       0:07:24.186 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1",
        "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "xfs",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "cryptsetup",
        "lvm2"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                    "_kernel_device": "/dev/dm-1",
                    "_mount_id": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": true,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": 0,
                    "encryption_luks_version": "luks2",
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "xfs",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Monday 20 January 2025  06:18:59 -0500 (0:00:04.280)       0:07:28.466 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Monday 20 January 2025  06:18:59 -0500 (0:00:00.043)       0:07:28.510 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371904.0871952,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "292c7f1e5b04e6e995ce2dec9cf6e125e041b744",
        "ctime": 1737371904.0841951,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 207618244,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1737371904.0841951,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1425,
        "uid": 0,
        "version": "261612921",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Monday 20 January 2025  06:19:00 -0500 (0:00:00.453)       0:07:28.964 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:19:00 -0500 (0:00:00.023)       0:07:28.987 ******** 
changed: [managed-node1] => (item=systemd-cryptsetup@luks\x2d4fd96f37\x2dc9c1\x2d41aa\x2dbce3\x2d48eefe3c9f81.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luks\\x2d4fd96f37\\x2dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
    "name": "systemd-cryptsetup@luks\\x2d4fd96f37\\x2dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luks\\x2d4fd96f37\\x2dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/dev/null",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luks\\x2d4fd96f37\\x2dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luks\\x2d4fd96f37\\x2dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service is masked.\"",
        "LoadState": "masked",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luks\\x2d4fd96f37\\x2dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFileState": "masked",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}
changed: [managed-node1] => (item=systemd-cryptsetup@luk...dc9c1\x2d41aa\x2dbce3\x2d48eefe3c9f81.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luk...dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
    "name": "systemd-cryptsetup@luk...dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luk...dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/dev/null",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luk...dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luk...dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service is masked.\"",
        "LoadState": "masked",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luk...dc9c1\\x2d41aa\\x2dbce3\\x2d48eefe3c9f81.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFileState": "masked",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Monday 20 January 2025  06:19:01 -0500 (0:00:01.385)       0:07:30.372 ******** 
ok: [managed-node1] => {
    "blivet_output": {
        "actions": [],
        "changed": false,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1",
            "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "xfs",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "cryptsetup",
            "lvm2"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                        "_kernel_device": "/dev/dm-1",
                        "_mount_id": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": true,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": 0,
                        "encryption_luks_version": "luks2",
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "xfs",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Monday 20 January 2025  06:19:01 -0500 (0:00:00.040)       0:07:30.413 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                        "_kernel_device": "/dev/dm-1",
                        "_mount_id": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": true,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": 0,
                        "encryption_luks_version": "luks2",
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "xfs",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Monday 20 January 2025  06:19:01 -0500 (0:00:00.037)       0:07:30.450 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Monday 20 January 2025  06:19:01 -0500 (0:00:00.034)       0:07:30.485 ******** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Monday 20 January 2025  06:19:01 -0500 (0:00:00.047)       0:07:30.533 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Monday 20 January 2025  06:19:02 -0500 (0:00:00.691)       0:07:31.224 ******** 
ok: [managed-node1] => (item={'src': '/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Monday 20 January 2025  06:19:02 -0500 (0:00:00.442)       0:07:31.666 ******** 
skipping: [managed-node1] => (item={'src': '/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Monday 20 January 2025  06:19:02 -0500 (0:00:00.042)       0:07:31.709 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Monday 20 January 2025  06:19:03 -0500 (0:00:00.683)       0:07:32.393 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371908.179205,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "d56e4da6f74b56543030966b4dd01cc882304a63",
        "ctime": 1737371906.2752004,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 446693515,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0600",
        "mtime": 1737371906.2752004,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 66,
        "uid": 0,
        "version": "252020374",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Monday 20 January 2025  06:19:04 -0500 (0:00:00.429)       0:07:32.822 ******** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Monday 20 January 2025  06:19:04 -0500 (0:00:00.029)       0:07:32.851 ******** 
ok: [managed-node1]

TASK [Assert preservation of encryption settings on existing LVM volume] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:423
Monday 20 January 2025  06:19:04 -0500 (0:00:00.911)       0:07:33.763 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify role results] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:430
Monday 20 January 2025  06:19:05 -0500 (0:00:00.046)       0:07:33.809 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node1

TASK [Print out pool information] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Monday 20 January 2025  06:19:05 -0500 (0:00:00.060)       0:07:33.870 ******** 
ok: [managed-node1] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                    "_kernel_device": "/dev/dm-1",
                    "_mount_id": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": true,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": 0,
                    "encryption_luks_version": "luks2",
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "xfs",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Monday 20 January 2025  06:19:05 -0500 (0:00:00.043)       0:07:33.914 ******** 
skipping: [managed-node1] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Monday 20 January 2025  06:19:05 -0500 (0:00:00.031)       0:07:33.945 ******** 
ok: [managed-node1] => {
    "changed": false,
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "crypto_LUKS",
            "label": "",
            "mountpoint": "",
            "name": "/dev/mapper/foo-test1",
            "size": "4G",
            "type": "lvm",
            "uuid": "940b35f5-2180-4a53-8562-e5b2ccb3efe6"
        },
        "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
            "size": "4G",
            "type": "crypt",
            "uuid": "9d7054c7-e815-4da1-9cee-d4e79336c702"
        },
        "/dev/sda": {
            "fstype": "LVM2_member",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "kSV36d-iWVA-92sg-L6wf-6hE7-Eik8-GLOUAJ"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Monday 20 January 2025  06:19:05 -0500 (0:00:00.442)       0:07:34.387 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.002683",
    "end": "2025-01-20 06:19:05.959498",
    "rc": 0,
    "start": "2025-01-20 06:19:05.956815"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6 /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Monday 20 January 2025  06:19:06 -0500 (0:00:00.404)       0:07:34.792 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002651",
    "end": "2025-01-20 06:19:06.464818",
    "failed_when_result": false,
    "rc": 0,
    "start": "2025-01-20 06:19:06.462167"
}

STDOUT:

luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6 /dev/mapper/foo-test1 -

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Monday 20 January 2025  06:19:06 -0500 (0:00:00.493)       0:07:35.285 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node1

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Monday 20 January 2025  06:19:06 -0500 (0:00:00.042)       0:07:35.327 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Monday 20 January 2025  06:19:06 -0500 (0:00:00.020)       0:07:35.348 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "vgs",
        "--noheadings",
        "--binary",
        "-o",
        "shared",
        "foo"
    ],
    "delta": "0:00:00.029331",
    "end": "2025-01-20 06:19:06.936554",
    "rc": 0,
    "start": "2025-01-20 06:19:06.907223"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Monday 20 January 2025  06:19:07 -0500 (0:00:00.472)       0:07:35.820 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Monday 20 January 2025  06:19:07 -0500 (0:00:00.043)       0:07:35.864 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node1

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Monday 20 January 2025  06:19:07 -0500 (0:00:00.155)       0:07:36.019 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Monday 20 January 2025  06:19:07 -0500 (0:00:00.039)       0:07:36.059 ******** 
ok: [managed-node1] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sda",
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Monday 20 January 2025  06:19:07 -0500 (0:00:00.381)       0:07:36.440 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Monday 20 January 2025  06:19:07 -0500 (0:00:00.025)       0:07:36.466 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Monday 20 January 2025  06:19:07 -0500 (0:00:00.026)       0:07:36.492 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Monday 20 January 2025  06:19:07 -0500 (0:00:00.027)       0:07:36.519 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Monday 20 January 2025  06:19:07 -0500 (0:00:00.024)       0:07:36.544 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Monday 20 January 2025  06:19:07 -0500 (0:00:00.024)       0:07:36.568 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Monday 20 January 2025  06:19:07 -0500 (0:00:00.019)       0:07:36.588 ******** 
ok: [managed-node1] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Monday 20 January 2025  06:19:07 -0500 (0:00:00.033)       0:07:36.621 ******** 
ok: [managed-node1] => {
    "changed": false,
    "failed_when_result": false,
    "rc": 1
}

STDERR:

Shared connection to 10.31.8.138 closed.



MSG:

non-zero return code

TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:74
Monday 20 January 2025  06:19:08 -0500 (0:00:00.423)       0:07:37.045 ******** 
skipping: [managed-node1] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:84
Monday 20 January 2025  06:19:08 -0500 (0:00:00.029)       0:07:37.075 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node1

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Monday 20 January 2025  06:19:08 -0500 (0:00:00.047)       0:07:37.123 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Monday 20 January 2025  06:19:08 -0500 (0:00:00.023)       0:07:37.146 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Monday 20 January 2025  06:19:08 -0500 (0:00:00.024)       0:07:37.171 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Monday 20 January 2025  06:19:08 -0500 (0:00:00.022)       0:07:37.194 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Monday 20 January 2025  06:19:08 -0500 (0:00:00.024)       0:07:37.218 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Monday 20 January 2025  06:19:08 -0500 (0:00:00.021)       0:07:37.239 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Monday 20 January 2025  06:19:08 -0500 (0:00:00.019)       0:07:37.258 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Monday 20 January 2025  06:19:08 -0500 (0:00:00.021)       0:07:37.280 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Monday 20 January 2025  06:19:08 -0500 (0:00:00.020)       0:07:37.300 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Monday 20 January 2025  06:19:08 -0500 (0:00:00.019)       0:07:37.320 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Monday 20 January 2025  06:19:08 -0500 (0:00:00.019)       0:07:37.339 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:87
Monday 20 January 2025  06:19:08 -0500 (0:00:00.019)       0:07:37.359 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node1

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Monday 20 January 2025  06:19:08 -0500 (0:00:00.047)       0:07:37.406 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node1

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Monday 20 January 2025  06:19:08 -0500 (0:00:00.056)       0:07:37.463 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Monday 20 January 2025  06:19:08 -0500 (0:00:00.022)       0:07:37.485 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Monday 20 January 2025  06:19:08 -0500 (0:00:00.024)       0:07:37.509 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Monday 20 January 2025  06:19:08 -0500 (0:00:00.025)       0:07:37.535 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Monday 20 January 2025  06:19:08 -0500 (0:00:00.021)       0:07:37.557 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Monday 20 January 2025  06:19:08 -0500 (0:00:00.027)       0:07:37.584 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Monday 20 January 2025  06:19:08 -0500 (0:00:00.022)       0:07:37.606 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:90
Monday 20 January 2025  06:19:08 -0500 (0:00:00.023)       0:07:37.629 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node1

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Monday 20 January 2025  06:19:08 -0500 (0:00:00.043)       0:07:37.673 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node1

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Monday 20 January 2025  06:19:08 -0500 (0:00:00.039)       0:07:37.713 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Monday 20 January 2025  06:19:08 -0500 (0:00:00.020)       0:07:37.734 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Monday 20 January 2025  06:19:08 -0500 (0:00:00.019)       0:07:37.753 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Monday 20 January 2025  06:19:09 -0500 (0:00:00.019)       0:07:37.773 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:93
Monday 20 January 2025  06:19:09 -0500 (0:00:00.025)       0:07:37.798 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node1

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Monday 20 January 2025  06:19:09 -0500 (0:00:00.079)       0:07:37.878 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Monday 20 January 2025  06:19:09 -0500 (0:00:00.067)       0:07:37.945 ******** 
skipping: [managed-node1] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Monday 20 January 2025  06:19:09 -0500 (0:00:00.026)       0:07:37.971 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node1

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Monday 20 January 2025  06:19:09 -0500 (0:00:00.048)       0:07:38.020 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Monday 20 January 2025  06:19:09 -0500 (0:00:00.027)       0:07:38.048 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Monday 20 January 2025  06:19:09 -0500 (0:00:00.026)       0:07:38.074 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Monday 20 January 2025  06:19:09 -0500 (0:00:00.020)       0:07:38.094 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Monday 20 January 2025  06:19:09 -0500 (0:00:00.019)       0:07:38.114 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Monday 20 January 2025  06:19:09 -0500 (0:00:00.020)       0:07:38.134 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Monday 20 January 2025  06:19:09 -0500 (0:00:00.018)       0:07:38.153 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:96
Monday 20 January 2025  06:19:09 -0500 (0:00:00.019)       0:07:38.172 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node1

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Monday 20 January 2025  06:19:09 -0500 (0:00:00.050)       0:07:38.222 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node1

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Monday 20 January 2025  06:19:09 -0500 (0:00:00.054)       0:07:38.277 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Monday 20 January 2025  06:19:09 -0500 (0:00:00.032)       0:07:38.310 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Monday 20 January 2025  06:19:09 -0500 (0:00:00.038)       0:07:38.349 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Monday 20 January 2025  06:19:09 -0500 (0:00:00.031)       0:07:38.380 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Monday 20 January 2025  06:19:09 -0500 (0:00:00.021)       0:07:38.401 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Monday 20 January 2025  06:19:09 -0500 (0:00:00.026)       0:07:38.427 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Monday 20 January 2025  06:19:09 -0500 (0:00:00.024)       0:07:38.452 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:99
Monday 20 January 2025  06:19:09 -0500 (0:00:00.022)       0:07:38.475 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node1

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Monday 20 January 2025  06:19:09 -0500 (0:00:00.058)       0:07:38.534 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Monday 20 January 2025  06:19:09 -0500 (0:00:00.026)       0:07:38.560 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Monday 20 January 2025  06:19:09 -0500 (0:00:00.022)       0:07:38.582 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Monday 20 January 2025  06:19:09 -0500 (0:00:00.020)       0:07:38.603 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Monday 20 January 2025  06:19:09 -0500 (0:00:00.019)       0:07:38.622 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Monday 20 January 2025  06:19:09 -0500 (0:00:00.020)       0:07:38.643 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:102
Monday 20 January 2025  06:19:09 -0500 (0:00:00.019)       0:07:38.663 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Monday 20 January 2025  06:19:09 -0500 (0:00:00.019)       0:07:38.682 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node1

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Monday 20 January 2025  06:19:09 -0500 (0:00:00.043)       0:07:38.726 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Monday 20 January 2025  06:19:09 -0500 (0:00:00.035)       0:07:38.762 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node1

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Monday 20 January 2025  06:19:10 -0500 (0:00:00.205)       0:07:38.968 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Monday 20 January 2025  06:19:10 -0500 (0:00:00.088)       0:07:39.057 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Monday 20 January 2025  06:19:10 -0500 (0:00:00.056)       0:07:39.114 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Monday 20 January 2025  06:19:10 -0500 (0:00:00.062)       0:07:39.177 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Monday 20 January 2025  06:19:10 -0500 (0:00:00.054)       0:07:39.231 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Monday 20 January 2025  06:19:10 -0500 (0:00:00.040)       0:07:39.271 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Monday 20 January 2025  06:19:10 -0500 (0:00:00.053)       0:07:39.325 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Monday 20 January 2025  06:19:10 -0500 (0:00:00.044)       0:07:39.369 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Monday 20 January 2025  06:19:10 -0500 (0:00:00.030)       0:07:39.400 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Monday 20 January 2025  06:19:10 -0500 (0:00:00.030)       0:07:39.431 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Monday 20 January 2025  06:19:10 -0500 (0:00:00.027)       0:07:39.459 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Monday 20 January 2025  06:19:10 -0500 (0:00:00.028)       0:07:39.488 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Monday 20 January 2025  06:19:10 -0500 (0:00:00.068)       0:07:39.556 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Monday 20 January 2025  06:19:10 -0500 (0:00:00.073)       0:07:39.630 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Monday 20 January 2025  06:19:10 -0500 (0:00:00.058)       0:07:39.689 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Monday 20 January 2025  06:19:10 -0500 (0:00:00.075)       0:07:39.764 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Monday 20 January 2025  06:19:11 -0500 (0:00:00.044)       0:07:39.809 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Monday 20 January 2025  06:19:11 -0500 (0:00:00.030)       0:07:39.840 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Monday 20 January 2025  06:19:11 -0500 (0:00:00.078)       0:07:39.919 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Monday 20 January 2025  06:19:11 -0500 (0:00:00.109)       0:07:40.028 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371917.996229,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1737371901.3091884,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 671645,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1737371901.3091884,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Monday 20 January 2025  06:19:11 -0500 (0:00:00.497)       0:07:40.526 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Monday 20 January 2025  06:19:11 -0500 (0:00:00.044)       0:07:40.570 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Monday 20 January 2025  06:19:11 -0500 (0:00:00.049)       0:07:40.620 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Monday 20 January 2025  06:19:11 -0500 (0:00:00.038)       0:07:40.659 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Monday 20 January 2025  06:19:11 -0500 (0:00:00.052)       0:07:40.712 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Monday 20 January 2025  06:19:11 -0500 (0:00:00.031)       0:07:40.743 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Monday 20 January 2025  06:19:12 -0500 (0:00:00.044)       0:07:40.788 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371939.5492811,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1737371901.4491887,
        "dev": 6,
        "device_type": 64769,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 672782,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1737371901.4491887,
        "nlink": 1,
        "path": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Monday 20 January 2025  06:19:12 -0500 (0:00:00.596)       0:07:41.384 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Monday 20 January 2025  06:19:15 -0500 (0:00:03.089)       0:07:44.474 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cryptsetup",
        "luksDump",
        "/dev/mapper/foo-test1"
    ],
    "delta": "0:00:00.011308",
    "end": "2025-01-20 06:19:16.083321",
    "rc": 0,
    "start": "2025-01-20 06:19:16.072013"
}

STDOUT:

LUKS header information
Version:       	2
Epoch:         	3
Metadata area: 	16384 [bytes]
Keyslots area: 	16744448 [bytes]
UUID:          	940b35f5-2180-4a53-8562-e5b2ccb3efe6
Label:         	(no label)
Subsystem:     	(no subsystem)
Flags:       	(no flags)

Data segments:
  0: crypt
	offset: 16777216 [bytes]
	length: (whole device)
	cipher: aes-xts-plain64
	sector: 512 [bytes]

Keyslots:
  0: luks2
	Key:        512 bits
	Priority:   normal
	Cipher:     aes-xts-plain64
	Cipher key: 512 bits
	PBKDF:      argon2i
	Time cost:  4
	Memory:     786697
	Threads:    2
	Salt:       91 2f 2a 22 b6 0e 7a bd 9f 87 7d 5d 23 fe ac 1b 
	            ad 4f fd 0b 88 3c ac d7 3e 11 7a 92 95 fd 43 8d 
	AF stripes: 4000
	AF hash:    sha256
	Area offset:32768 [bytes]
	Area length:258048 [bytes]
	Digest ID:  0
Tokens:
Digests:
  0: pbkdf2
	Hash:       sha256
	Iterations: 120249
	Salt:       7d 7b e3 a9 e1 08 76 be 61 e0 d0 f1 8e fd f4 89 
	            ed 05 7e c2 0d a2 41 7a 7a 49 f0 17 3b 68 5c 74 
	Digest:     7d 70 07 42 f6 fe 4f 52 91 9e 37 e3 42 46 62 2a 
	            fa ec e0 d3 fa 94 bf 70 90 4e de db f9 af e1 c7 

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Monday 20 January 2025  06:19:16 -0500 (0:00:00.452)       0:07:44.926 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Monday 20 January 2025  06:19:16 -0500 (0:00:00.044)       0:07:44.971 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Monday 20 January 2025  06:19:16 -0500 (0:00:00.045)       0:07:45.017 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Monday 20 January 2025  06:19:16 -0500 (0:00:00.039)       0:07:45.056 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Monday 20 January 2025  06:19:16 -0500 (0:00:00.039)       0:07:45.096 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Monday 20 January 2025  06:19:16 -0500 (0:00:00.046)       0:07:45.142 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Monday 20 January 2025  06:19:16 -0500 (0:00:00.033)       0:07:45.175 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Monday 20 January 2025  06:19:16 -0500 (0:00:00.032)       0:07:45.208 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [
            "luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6 /dev/mapper/foo-test1 -"
        ],
        "_storage_test_expected_crypttab_entries": "1",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Monday 20 January 2025  06:19:16 -0500 (0:00:00.046)       0:07:45.254 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Monday 20 January 2025  06:19:16 -0500 (0:00:00.045)       0:07:45.300 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Monday 20 January 2025  06:19:16 -0500 (0:00:00.046)       0:07:45.347 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Monday 20 January 2025  06:19:16 -0500 (0:00:00.040)       0:07:45.387 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Monday 20 January 2025  06:19:16 -0500 (0:00:00.042)       0:07:45.429 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Monday 20 January 2025  06:19:16 -0500 (0:00:00.030)       0:07:45.460 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Monday 20 January 2025  06:19:16 -0500 (0:00:00.023)       0:07:45.483 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Monday 20 January 2025  06:19:16 -0500 (0:00:00.024)       0:07:45.507 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Monday 20 January 2025  06:19:16 -0500 (0:00:00.019)       0:07:45.527 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Monday 20 January 2025  06:19:16 -0500 (0:00:00.026)       0:07:45.554 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Monday 20 January 2025  06:19:16 -0500 (0:00:00.021)       0:07:45.575 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Monday 20 January 2025  06:19:16 -0500 (0:00:00.027)       0:07:45.602 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Monday 20 January 2025  06:19:16 -0500 (0:00:00.023)       0:07:45.626 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Monday 20 January 2025  06:19:16 -0500 (0:00:00.025)       0:07:45.652 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Monday 20 January 2025  06:19:16 -0500 (0:00:00.020)       0:07:45.673 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Monday 20 January 2025  06:19:16 -0500 (0:00:00.021)       0:07:45.694 ******** 
ok: [managed-node1] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Monday 20 January 2025  06:19:17 -0500 (0:00:00.355)       0:07:46.050 ******** 
ok: [managed-node1] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Monday 20 January 2025  06:19:17 -0500 (0:00:00.355)       0:07:46.406 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_expected_size": "4294967296"
    },
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Monday 20 January 2025  06:19:17 -0500 (0:00:00.029)       0:07:46.435 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Monday 20 January 2025  06:19:17 -0500 (0:00:00.023)       0:07:46.458 ******** 
ok: [managed-node1] => {
    "bytes": 10726680821,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Monday 20 January 2025  06:19:18 -0500 (0:00:00.364)       0:07:46.823 ******** 
skipping: [managed-node1] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Monday 20 January 2025  06:19:18 -0500 (0:00:00.026)       0:07:46.849 ******** 
skipping: [managed-node1] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Monday 20 January 2025  06:19:18 -0500 (0:00:00.023)       0:07:46.873 ******** 
skipping: [managed-node1] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Monday 20 January 2025  06:19:18 -0500 (0:00:00.023)       0:07:46.897 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Monday 20 January 2025  06:19:18 -0500 (0:00:00.023)       0:07:46.920 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Monday 20 January 2025  06:19:18 -0500 (0:00:00.022)       0:07:46.942 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Monday 20 January 2025  06:19:18 -0500 (0:00:00.019)       0:07:46.962 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Monday 20 January 2025  06:19:18 -0500 (0:00:00.019)       0:07:46.981 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Monday 20 January 2025  06:19:18 -0500 (0:00:00.020)       0:07:47.002 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Monday 20 January 2025  06:19:18 -0500 (0:00:00.018)       0:07:47.021 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Monday 20 January 2025  06:19:18 -0500 (0:00:00.019)       0:07:47.041 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Monday 20 January 2025  06:19:18 -0500 (0:00:00.021)       0:07:47.063 ******** 
skipping: [managed-node1] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Monday 20 January 2025  06:19:18 -0500 (0:00:00.019)       0:07:47.082 ******** 
skipping: [managed-node1] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Monday 20 January 2025  06:19:18 -0500 (0:00:00.020)       0:07:47.103 ******** 
skipping: [managed-node1] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Monday 20 January 2025  06:19:18 -0500 (0:00:00.020)       0:07:47.123 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Monday 20 January 2025  06:19:18 -0500 (0:00:00.019)       0:07:47.143 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Monday 20 January 2025  06:19:18 -0500 (0:00:00.019)       0:07:47.163 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Monday 20 January 2025  06:19:18 -0500 (0:00:00.021)       0:07:47.185 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Monday 20 January 2025  06:19:18 -0500 (0:00:00.020)       0:07:47.206 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Monday 20 January 2025  06:19:18 -0500 (0:00:00.019)       0:07:47.225 ******** 
ok: [managed-node1] => {
    "storage_test_actual_size": {
        "bytes": 4294967296,
        "changed": false,
        "failed": false,
        "lvm": "4g",
        "parted": "4GiB",
        "size": "4 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Monday 20 January 2025  06:19:18 -0500 (0:00:00.023)       0:07:47.248 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Monday 20 January 2025  06:19:18 -0500 (0:00:00.023)       0:07:47.272 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Monday 20 January 2025  06:19:18 -0500 (0:00:00.029)       0:07:47.301 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.023633",
    "end": "2025-01-20 06:19:18.867464",
    "rc": 0,
    "start": "2025-01-20 06:19:18.843831"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Monday 20 January 2025  06:19:18 -0500 (0:00:00.405)       0:07:47.707 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Monday 20 January 2025  06:19:19 -0500 (0:00:00.096)       0:07:47.803 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Monday 20 January 2025  06:19:19 -0500 (0:00:00.046)       0:07:47.850 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Monday 20 January 2025  06:19:19 -0500 (0:00:00.039)       0:07:47.889 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Monday 20 January 2025  06:19:19 -0500 (0:00:00.040)       0:07:47.929 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Monday 20 January 2025  06:19:19 -0500 (0:00:00.036)       0:07:47.966 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Monday 20 January 2025  06:19:19 -0500 (0:00:00.035)       0:07:48.001 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Monday 20 January 2025  06:19:19 -0500 (0:00:00.034)       0:07:48.035 ******** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Monday 20 January 2025  06:19:19 -0500 (0:00:00.039)       0:07:48.075 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Create a file] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml:12
Monday 20 January 2025  06:19:19 -0500 (0:00:00.023)       0:07:48.099 ******** 
changed: [managed-node1] => {
    "changed": true,
    "dest": "/opt/test1/quux",
    "gid": 0,
    "group": "root",
    "mode": "0644",
    "owner": "root",
    "secontext": "unconfined_u:object_r:unlabeled_t:s0",
    "size": 0,
    "state": "file",
    "uid": 0
}

TASK [Test for correct handling of safe_mode] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:436
Monday 20 January 2025  06:19:19 -0500 (0:00:00.430)       0:07:48.530 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node1

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Monday 20 January 2025  06:19:19 -0500 (0:00:00.072)       0:07:48.602 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_pools_global": [],
        "storage_safe_mode_global": true,
        "storage_volumes_global": []
    },
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Monday 20 January 2025  06:19:19 -0500 (0:00:00.040)       0:07:48.643 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:19:19 -0500 (0:00:00.063)       0:07:48.706 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:19:19 -0500 (0:00:00.045)       0:07:48.752 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:19:20 -0500 (0:00:00.037)       0:07:48.790 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:19:20 -0500 (0:00:00.087)       0:07:48.878 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:19:20 -0500 (0:00:00.030)       0:07:48.908 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:19:20 -0500 (0:00:00.033)       0:07:48.942 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:19:20 -0500 (0:00:00.029)       0:07:48.971 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:19:20 -0500 (0:00:00.026)       0:07:48.997 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:19:20 -0500 (0:00:00.106)       0:07:49.104 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:19:23 -0500 (0:00:03.011)       0:07:52.116 ******** 
ok: [managed-node1] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "encryption": false,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "yabbadabbadoo",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:19:23 -0500 (0:00:00.064)       0:07:52.180 ******** 
ok: [managed-node1] => {
    "storage_volumes": []
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:19:23 -0500 (0:00:00.066)       0:07:52.246 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "lvm2"
    ],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:19:28 -0500 (0:00:04.530)       0:07:56.776 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:19:28 -0500 (0:00:00.113)       0:07:56.890 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:19:28 -0500 (0:00:00.040)       0:07:56.930 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:19:28 -0500 (0:00:00.048)       0:07:56.979 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:19:28 -0500 (0:00:00.041)       0:07:57.021 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:19:31 -0500 (0:00:02.986)       0:08:00.007 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service": {
                "name": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
                "source": "systemd",
                "state": "inactive",
                "status": "generated"
            },
            "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service": {
                "name": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:19:33 -0500 (0:00:01.851)       0:08:01.858 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": [
            "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
            "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service"
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:19:33 -0500 (0:00:00.033)       0:08:01.892 ******** 
changed: [managed-node1] => (item=systemd-cryptsetup@luks\x2d940b35f5\x2d2180\x2d4a53\x2d8562\x2de5b2ccb3efe6.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "name": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "After": "systemd-journald.socket system-systemd\\x2dcryptsetup.slice cryptsetup-pre.target dev-mapper-foo\\x2dtest1.device",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "Before": "umount.target cryptsetup.target",
        "BindsTo": "dev-mapper-foo\\x2dtest1.device",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "yes",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "Conflicts": "umount.target",
        "ControlPID": "0",
        "DefaultDependencies": "no",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "Cryptography Setup for luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
        "DevicePolicy": "auto",
        "Documentation": "man:crypttab(5) man:systemd-cryptsetup-generator(8) man:systemd-cryptsetup@.service(8)",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6 /dev/mapper/foo-test1 -  ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
        "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "IgnoreOnIsolate": "yes",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "shared",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "0",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "65536",
        "LimitMEMLOCKSoft": "65536",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "262144",
        "LimitNOFILESoft": "1024",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadState": "loaded",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "yes",
        "RemoveIPC": "no",
        "RequiredBy": "dev-mapper-luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.device cryptsetup.target",
        "Requires": "system-systemd\\x2dcryptsetup.slice",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "Slice": "system-systemd\\x2dcryptsetup.slice",
        "SourcePath": "/etc/crypttab",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "journal",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestamp": "Mon 2025-01-20 06:18:55 EST",
        "StateChangeTimestampMonotonic": "2753199841",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "infinity",
        "TimeoutStopUSec": "infinity",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "Type": "oneshot",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFilePreset": "disabled",
        "UnitFileState": "generated",
        "UtmpMode": "init",
        "WantedBy": "dev-mapper-foo\\x2dtest1.device",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}
changed: [managed-node1] => (item=systemd-cryptsetup@luk...d2180\x2d4a53\x2d8562\x2de5b2ccb3efe6.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "name": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.NoSuchUnit \"Unit systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service not found.\"",
        "LoadState": "not-found",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:19:34 -0500 (0:00:01.376)       0:08:03.269 ******** 
fatal: [managed-node1]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

cannot remove existing formatting on device 'luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6' in safe mode due to encryption removal

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Monday 20 January 2025  06:19:38 -0500 (0:00:04.061)       0:08:07.330 ******** 
fatal: [managed-node1]: FAILED! => {
    "changed": false
}

MSG:

{'msg': "cannot remove existing formatting on device 'luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6' in safe mode due to encryption removal", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'encryption_clevis_pin': None, 'encryption_tang_url': None, 'encryption_tang_thumbprint': None, 'grow_to_fill': False, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'shared': False, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': 0, 'encryption_luks_version': 'luks2', 'encryption_password': 'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}]}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'grow_to_fill': False, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, 'shared': False}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:19:38 -0500 (0:00:00.029)       0:08:07.360 ******** 
changed: [managed-node1] => (item=systemd-cryptsetup@luks\x2d940b35f5\x2d2180\x2d4a53\x2d8562\x2de5b2ccb3efe6.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "name": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/dev/null",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service is masked.\"",
        "LoadState": "masked",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "RequiredBy": "dev-mapper-luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.device cryptsetup.target",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestamp": "Mon 2025-01-20 06:18:55 EST",
        "StateChangeTimestampMonotonic": "2753199841",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFileState": "masked",
        "UtmpMode": "init",
        "WantedBy": "dev-mapper-foo\\x2dtest1.device",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}
changed: [managed-node1] => (item=systemd-cryptsetup@luk...d2180\x2d4a53\x2d8562\x2de5b2ccb3efe6.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "name": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/dev/null",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service is masked.\"",
        "LoadState": "masked",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFileState": "masked",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Monday 20 January 2025  06:19:39 -0500 (0:00:01.378)       0:08:08.738 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Monday 20 January 2025  06:19:40 -0500 (0:00:00.036)       0:08:08.774 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Monday 20 January 2025  06:19:40 -0500 (0:00:00.038)       0:08:08.813 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Stat the file] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:11
Monday 20 January 2025  06:19:40 -0500 (0:00:00.032)       0:08:08.846 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371959.68133,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1737371959.68133,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 131,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0644",
        "mtime": 1737371959.68133,
        "nlink": 1,
        "path": "/opt/test1/quux",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "4110484169",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Assert file presence] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:16
Monday 20 January 2025  06:19:40 -0500 (0:00:00.382)       0:08:09.228 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Remove the encryption layer] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:460
Monday 20 January 2025  06:19:40 -0500 (0:00:00.036)       0:08:09.265 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:19:40 -0500 (0:00:00.070)       0:08:09.335 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:19:40 -0500 (0:00:00.042)       0:08:09.378 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:19:40 -0500 (0:00:00.028)       0:08:09.406 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:19:40 -0500 (0:00:00.055)       0:08:09.462 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:19:40 -0500 (0:00:00.020)       0:08:09.482 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:19:40 -0500 (0:00:00.021)       0:08:09.504 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:19:40 -0500 (0:00:00.019)       0:08:09.524 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:19:40 -0500 (0:00:00.021)       0:08:09.545 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:19:40 -0500 (0:00:00.068)       0:08:09.614 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:19:43 -0500 (0:00:03.082)       0:08:12.696 ******** 
ok: [managed-node1] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "encryption": false,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "yabbadabbadoo",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:19:43 -0500 (0:00:00.027)       0:08:12.724 ******** 
ok: [managed-node1] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:19:43 -0500 (0:00:00.024)       0:08:12.749 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "lvm2"
    ],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:19:48 -0500 (0:00:04.321)       0:08:17.070 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:19:48 -0500 (0:00:00.067)       0:08:17.138 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:19:48 -0500 (0:00:00.035)       0:08:17.174 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:19:48 -0500 (0:00:00.035)       0:08:17.209 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:19:48 -0500 (0:00:00.085)       0:08:17.295 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:19:51 -0500 (0:00:02.965)       0:08:20.260 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service": {
                "name": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
                "source": "systemd",
                "state": "inactive",
                "status": "generated"
            },
            "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service": {
                "name": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:19:53 -0500 (0:00:01.965)       0:08:22.225 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": [
            "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
            "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service"
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:19:53 -0500 (0:00:00.079)       0:08:22.305 ******** 
changed: [managed-node1] => (item=systemd-cryptsetup@luks\x2d940b35f5\x2d2180\x2d4a53\x2d8562\x2de5b2ccb3efe6.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "name": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "After": "systemd-journald.socket system-systemd\\x2dcryptsetup.slice cryptsetup-pre.target dev-mapper-foo\\x2dtest1.device",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "Before": "umount.target cryptsetup.target",
        "BindsTo": "dev-mapper-foo\\x2dtest1.device",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "yes",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "Conflicts": "umount.target",
        "ControlPID": "0",
        "DefaultDependencies": "no",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "Cryptography Setup for luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
        "DevicePolicy": "auto",
        "Documentation": "man:crypttab(5) man:systemd-cryptsetup-generator(8) man:systemd-cryptsetup@.service(8)",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6 /dev/mapper/foo-test1 -  ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
        "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "IgnoreOnIsolate": "yes",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "shared",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "0",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "65536",
        "LimitMEMLOCKSoft": "65536",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "262144",
        "LimitNOFILESoft": "1024",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadState": "loaded",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "yes",
        "RemoveIPC": "no",
        "RequiredBy": "dev-mapper-luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.device cryptsetup.target",
        "Requires": "system-systemd\\x2dcryptsetup.slice",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "Slice": "system-systemd\\x2dcryptsetup.slice",
        "SourcePath": "/etc/crypttab",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "journal",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestamp": "Mon 2025-01-20 06:18:55 EST",
        "StateChangeTimestampMonotonic": "2753199841",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "infinity",
        "TimeoutStopUSec": "infinity",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "Type": "oneshot",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFilePreset": "disabled",
        "UnitFileState": "generated",
        "UtmpMode": "init",
        "WantedBy": "dev-mapper-foo\\x2dtest1.device",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}
changed: [managed-node1] => (item=systemd-cryptsetup@luk...d2180\x2d4a53\x2d8562\x2de5b2ccb3efe6.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "name": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.NoSuchUnit \"Unit systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service not found.\"",
        "LoadState": "not-found",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:19:55 -0500 (0:00:01.787)       0:08:24.092 ******** 
changed: [managed-node1] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
            "fs_type": "xfs"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "luks"
        },
        {
            "action": "create format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "xfs"
        }
    ],
    "changed": true,
    "crypts": [
        {
            "backing_device": "/dev/mapper/foo-test1",
            "name": "luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
            "password": "-",
            "state": "absent"
        }
    ],
    "leaves": [
        "/dev/mapper/foo-test1",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1"
    ],
    "mounts": [
        {
            "fstype": "xfs",
            "path": "/opt/test1",
            "src": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
            "state": "absent"
        },
        {
            "dump": 0,
            "fstype": "xfs",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "lvm2"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": 0,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "xfs",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Monday 20 January 2025  06:20:00 -0500 (0:00:05.049)       0:08:29.142 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Monday 20 January 2025  06:20:00 -0500 (0:00:00.024)       0:08:29.167 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371904.0871952,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "292c7f1e5b04e6e995ce2dec9cf6e125e041b744",
        "ctime": 1737371904.0841951,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 207618244,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1737371904.0841951,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1425,
        "uid": 0,
        "version": "261612921",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Monday 20 January 2025  06:20:00 -0500 (0:00:00.385)       0:08:29.552 ******** 
ok: [managed-node1] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:20:01 -0500 (0:00:00.390)       0:08:29.942 ******** 
changed: [managed-node1] => (item=systemd-cryptsetup@luks\x2d940b35f5\x2d2180\x2d4a53\x2d8562\x2de5b2ccb3efe6.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "name": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/dev/null",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service is masked.\"",
        "LoadState": "masked",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "RequiredBy": "dev-mapper-luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.device cryptsetup.target",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestamp": "Mon 2025-01-20 06:18:55 EST",
        "StateChangeTimestampMonotonic": "2753199841",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFileState": "masked",
        "UtmpMode": "init",
        "WantedBy": "dev-mapper-foo\\x2dtest1.device",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}
changed: [managed-node1] => (item=systemd-cryptsetup@luk...d2180\x2d4a53\x2d8562\x2de5b2ccb3efe6.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "name": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/dev/null",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service is masked.\"",
        "LoadState": "masked",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFileState": "masked",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Monday 20 January 2025  06:20:02 -0500 (0:00:01.441)       0:08:31.384 ******** 
ok: [managed-node1] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                "fs_type": "xfs"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "luks"
            },
            {
                "action": "create format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "xfs"
            }
        ],
        "changed": true,
        "crypts": [
            {
                "backing_device": "/dev/mapper/foo-test1",
                "name": "luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                "password": "-",
                "state": "absent"
            }
        ],
        "failed": false,
        "leaves": [
            "/dev/mapper/foo-test1",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1"
        ],
        "mounts": [
            {
                "fstype": "xfs",
                "path": "/opt/test1",
                "src": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
                "state": "absent"
            },
            {
                "dump": 0,
                "fstype": "xfs",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "lvm2"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": 0,
                        "encryption_luks_version": "luks2",
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "xfs",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Monday 20 January 2025  06:20:02 -0500 (0:00:00.043)       0:08:31.428 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": 0,
                        "encryption_luks_version": "luks2",
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "xfs",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Monday 20 January 2025  06:20:02 -0500 (0:00:00.039)       0:08:31.468 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Monday 20 January 2025  06:20:02 -0500 (0:00:00.038)       0:08:31.507 ******** 
changed: [managed-node1] => (item={'src': '/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "fstype": "xfs",
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Monday 20 January 2025  06:20:03 -0500 (0:00:00.470)       0:08:31.977 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Monday 20 January 2025  06:20:03 -0500 (0:00:00.711)       0:08:32.689 ******** 
changed: [managed-node1] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Monday 20 January 2025  06:20:04 -0500 (0:00:00.421)       0:08:33.111 ******** 
skipping: [managed-node1] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Monday 20 January 2025  06:20:04 -0500 (0:00:00.045)       0:08:33.156 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Monday 20 January 2025  06:20:05 -0500 (0:00:00.640)       0:08:33.797 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737371908.179205,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "d56e4da6f74b56543030966b4dd01cc882304a63",
        "ctime": 1737371906.2752004,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 446693515,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0600",
        "mtime": 1737371906.2752004,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 66,
        "uid": 0,
        "version": "252020374",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Monday 20 January 2025  06:20:05 -0500 (0:00:00.457)       0:08:34.255 ******** 
changed: [managed-node1] => (item={'backing_device': '/dev/mapper/foo-test1', 'name': 'luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6', 'password': '-', 'state': 'absent'}) => {
    "ansible_loop_var": "entry",
    "backup": "",
    "changed": true,
    "entry": {
        "backing_device": "/dev/mapper/foo-test1",
        "name": "luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
        "password": "-",
        "state": "absent"
    },
    "found": 1
}

MSG:

1 line(s) removed

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Monday 20 January 2025  06:20:06 -0500 (0:00:00.734)       0:08:34.990 ******** 
ok: [managed-node1]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:477
Monday 20 January 2025  06:20:06 -0500 (0:00:00.759)       0:08:35.749 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node1

TASK [Print out pool information] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Monday 20 January 2025  06:20:07 -0500 (0:00:00.079)       0:08:35.828 ******** 
ok: [managed-node1] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": 0,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "xfs",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Monday 20 January 2025  06:20:07 -0500 (0:00:00.028)       0:08:35.857 ******** 
skipping: [managed-node1] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Monday 20 January 2025  06:20:07 -0500 (0:00:00.020)       0:08:35.878 ******** 
ok: [managed-node1] => {
    "changed": false,
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/foo-test1",
            "size": "4G",
            "type": "lvm",
            "uuid": "478b23b8-2d7b-45c3-90b1-d26f45155376"
        },
        "/dev/sda": {
            "fstype": "LVM2_member",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "kSV36d-iWVA-92sg-L6wf-6hE7-Eik8-GLOUAJ"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Monday 20 January 2025  06:20:07 -0500 (0:00:00.369)       0:08:36.247 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003233",
    "end": "2025-01-20 06:20:07.779491",
    "rc": 0,
    "start": "2025-01-20 06:20:07.776258"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Monday 20 January 2025  06:20:07 -0500 (0:00:00.355)       0:08:36.602 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002651",
    "end": "2025-01-20 06:20:08.134398",
    "failed_when_result": false,
    "rc": 0,
    "start": "2025-01-20 06:20:08.131747"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Monday 20 January 2025  06:20:08 -0500 (0:00:00.382)       0:08:36.984 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node1

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Monday 20 January 2025  06:20:08 -0500 (0:00:00.069)       0:08:37.053 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Monday 20 January 2025  06:20:08 -0500 (0:00:00.030)       0:08:37.084 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "vgs",
        "--noheadings",
        "--binary",
        "-o",
        "shared",
        "foo"
    ],
    "delta": "0:00:00.025144",
    "end": "2025-01-20 06:20:08.699465",
    "rc": 0,
    "start": "2025-01-20 06:20:08.674321"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Monday 20 January 2025  06:20:08 -0500 (0:00:00.446)       0:08:37.530 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Monday 20 January 2025  06:20:08 -0500 (0:00:00.045)       0:08:37.576 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node1

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Monday 20 January 2025  06:20:08 -0500 (0:00:00.073)       0:08:37.649 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Monday 20 January 2025  06:20:08 -0500 (0:00:00.042)       0:08:37.691 ******** 
ok: [managed-node1] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sda",
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Monday 20 January 2025  06:20:09 -0500 (0:00:00.414)       0:08:38.106 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Monday 20 January 2025  06:20:09 -0500 (0:00:00.039)       0:08:38.146 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Monday 20 January 2025  06:20:09 -0500 (0:00:00.040)       0:08:38.186 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Monday 20 January 2025  06:20:09 -0500 (0:00:00.044)       0:08:38.231 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Monday 20 January 2025  06:20:09 -0500 (0:00:00.042)       0:08:38.273 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Monday 20 January 2025  06:20:09 -0500 (0:00:00.051)       0:08:38.324 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Monday 20 January 2025  06:20:09 -0500 (0:00:00.057)       0:08:38.382 ******** 
ok: [managed-node1] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Monday 20 January 2025  06:20:09 -0500 (0:00:00.046)       0:08:38.428 ******** 
ok: [managed-node1] => {
    "changed": false,
    "failed_when_result": false,
    "rc": 1
}

STDERR:

Shared connection to 10.31.8.138 closed.



MSG:

non-zero return code

TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:74
Monday 20 January 2025  06:20:10 -0500 (0:00:00.650)       0:08:39.079 ******** 
skipping: [managed-node1] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:84
Monday 20 January 2025  06:20:10 -0500 (0:00:00.025)       0:08:39.105 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node1

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Monday 20 January 2025  06:20:10 -0500 (0:00:00.041)       0:08:39.146 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Monday 20 January 2025  06:20:10 -0500 (0:00:00.021)       0:08:39.167 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Monday 20 January 2025  06:20:10 -0500 (0:00:00.021)       0:08:39.188 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Monday 20 January 2025  06:20:10 -0500 (0:00:00.019)       0:08:39.208 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Monday 20 January 2025  06:20:10 -0500 (0:00:00.019)       0:08:39.228 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Monday 20 January 2025  06:20:10 -0500 (0:00:00.025)       0:08:39.253 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Monday 20 January 2025  06:20:10 -0500 (0:00:00.025)       0:08:39.279 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Monday 20 January 2025  06:20:10 -0500 (0:00:00.030)       0:08:39.309 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Monday 20 January 2025  06:20:10 -0500 (0:00:00.030)       0:08:39.339 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Monday 20 January 2025  06:20:10 -0500 (0:00:00.029)       0:08:39.368 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Monday 20 January 2025  06:20:10 -0500 (0:00:00.032)       0:08:39.400 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:87
Monday 20 January 2025  06:20:10 -0500 (0:00:00.025)       0:08:39.426 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node1

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Monday 20 January 2025  06:20:10 -0500 (0:00:00.063)       0:08:39.489 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node1

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Monday 20 January 2025  06:20:10 -0500 (0:00:00.064)       0:08:39.554 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Monday 20 January 2025  06:20:10 -0500 (0:00:00.031)       0:08:39.586 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Monday 20 January 2025  06:20:10 -0500 (0:00:00.072)       0:08:39.659 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Monday 20 January 2025  06:20:10 -0500 (0:00:00.029)       0:08:39.689 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Monday 20 January 2025  06:20:10 -0500 (0:00:00.027)       0:08:39.717 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Monday 20 January 2025  06:20:10 -0500 (0:00:00.025)       0:08:39.743 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Monday 20 January 2025  06:20:11 -0500 (0:00:00.025)       0:08:39.768 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:90
Monday 20 January 2025  06:20:11 -0500 (0:00:00.022)       0:08:39.790 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node1

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Monday 20 January 2025  06:20:11 -0500 (0:00:00.045)       0:08:39.836 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node1

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Monday 20 January 2025  06:20:11 -0500 (0:00:00.043)       0:08:39.880 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Monday 20 January 2025  06:20:11 -0500 (0:00:00.023)       0:08:39.903 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Monday 20 January 2025  06:20:11 -0500 (0:00:00.022)       0:08:39.925 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Monday 20 January 2025  06:20:11 -0500 (0:00:00.022)       0:08:39.948 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:93
Monday 20 January 2025  06:20:11 -0500 (0:00:00.021)       0:08:39.969 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node1

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Monday 20 January 2025  06:20:11 -0500 (0:00:00.048)       0:08:40.018 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Monday 20 January 2025  06:20:11 -0500 (0:00:00.038)       0:08:40.056 ******** 
skipping: [managed-node1] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Monday 20 January 2025  06:20:11 -0500 (0:00:00.029)       0:08:40.085 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node1

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Monday 20 January 2025  06:20:11 -0500 (0:00:00.040)       0:08:40.126 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Monday 20 January 2025  06:20:11 -0500 (0:00:00.026)       0:08:40.153 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Monday 20 January 2025  06:20:11 -0500 (0:00:00.029)       0:08:40.182 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Monday 20 January 2025  06:20:11 -0500 (0:00:00.022)       0:08:40.205 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Monday 20 January 2025  06:20:11 -0500 (0:00:00.021)       0:08:40.226 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Monday 20 January 2025  06:20:11 -0500 (0:00:00.021)       0:08:40.248 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Monday 20 January 2025  06:20:11 -0500 (0:00:00.020)       0:08:40.268 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:96
Monday 20 January 2025  06:20:11 -0500 (0:00:00.020)       0:08:40.288 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node1

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Monday 20 January 2025  06:20:11 -0500 (0:00:00.047)       0:08:40.336 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node1

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Monday 20 January 2025  06:20:11 -0500 (0:00:00.044)       0:08:40.381 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Monday 20 January 2025  06:20:11 -0500 (0:00:00.023)       0:08:40.405 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Monday 20 January 2025  06:20:11 -0500 (0:00:00.022)       0:08:40.427 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Monday 20 January 2025  06:20:11 -0500 (0:00:00.022)       0:08:40.450 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Monday 20 January 2025  06:20:11 -0500 (0:00:00.022)       0:08:40.472 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Monday 20 January 2025  06:20:11 -0500 (0:00:00.025)       0:08:40.497 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Monday 20 January 2025  06:20:11 -0500 (0:00:00.021)       0:08:40.519 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:99
Monday 20 January 2025  06:20:11 -0500 (0:00:00.023)       0:08:40.543 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node1

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Monday 20 January 2025  06:20:11 -0500 (0:00:00.050)       0:08:40.593 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Monday 20 January 2025  06:20:11 -0500 (0:00:00.022)       0:08:40.616 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Monday 20 January 2025  06:20:11 -0500 (0:00:00.024)       0:08:40.640 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Monday 20 January 2025  06:20:11 -0500 (0:00:00.022)       0:08:40.663 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Monday 20 January 2025  06:20:11 -0500 (0:00:00.021)       0:08:40.684 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Monday 20 January 2025  06:20:11 -0500 (0:00:00.021)       0:08:40.705 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:102
Monday 20 January 2025  06:20:11 -0500 (0:00:00.019)       0:08:40.725 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Monday 20 January 2025  06:20:11 -0500 (0:00:00.019)       0:08:40.745 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node1

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Monday 20 January 2025  06:20:12 -0500 (0:00:00.040)       0:08:40.785 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Monday 20 January 2025  06:20:12 -0500 (0:00:00.026)       0:08:40.812 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node1

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Monday 20 January 2025  06:20:12 -0500 (0:00:00.108)       0:08:40.920 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Monday 20 January 2025  06:20:12 -0500 (0:00:00.069)       0:08:40.990 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Monday 20 January 2025  06:20:12 -0500 (0:00:00.028)       0:08:41.018 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Monday 20 January 2025  06:20:12 -0500 (0:00:00.020)       0:08:41.039 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Monday 20 January 2025  06:20:12 -0500 (0:00:00.028)       0:08:41.067 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Monday 20 January 2025  06:20:12 -0500 (0:00:00.029)       0:08:41.097 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Monday 20 January 2025  06:20:12 -0500 (0:00:00.022)       0:08:41.119 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Monday 20 January 2025  06:20:12 -0500 (0:00:00.023)       0:08:41.142 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Monday 20 January 2025  06:20:12 -0500 (0:00:00.020)       0:08:41.163 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Monday 20 January 2025  06:20:12 -0500 (0:00:00.019)       0:08:41.183 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Monday 20 January 2025  06:20:12 -0500 (0:00:00.019)       0:08:41.202 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Monday 20 January 2025  06:20:12 -0500 (0:00:00.019)       0:08:41.221 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Monday 20 January 2025  06:20:12 -0500 (0:00:00.042)       0:08:41.264 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Monday 20 January 2025  06:20:12 -0500 (0:00:00.032)       0:08:41.296 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Monday 20 January 2025  06:20:12 -0500 (0:00:00.025)       0:08:41.322 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Monday 20 January 2025  06:20:12 -0500 (0:00:00.020)       0:08:41.343 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Monday 20 January 2025  06:20:12 -0500 (0:00:00.026)       0:08:41.369 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Monday 20 January 2025  06:20:12 -0500 (0:00:00.021)       0:08:41.390 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Monday 20 January 2025  06:20:12 -0500 (0:00:00.029)       0:08:41.420 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Monday 20 January 2025  06:20:12 -0500 (0:00:00.037)       0:08:41.457 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737372000.2604282,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1737372000.2604282,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 702121,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1737372000.2604282,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Monday 20 January 2025  06:20:13 -0500 (0:00:00.360)       0:08:41.818 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Monday 20 January 2025  06:20:13 -0500 (0:00:00.027)       0:08:41.846 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Monday 20 January 2025  06:20:13 -0500 (0:00:00.021)       0:08:41.868 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Monday 20 January 2025  06:20:13 -0500 (0:00:00.024)       0:08:41.892 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Monday 20 January 2025  06:20:13 -0500 (0:00:00.021)       0:08:41.914 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Monday 20 January 2025  06:20:13 -0500 (0:00:00.021)       0:08:41.936 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Monday 20 January 2025  06:20:13 -0500 (0:00:00.024)       0:08:41.960 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Monday 20 January 2025  06:20:13 -0500 (0:00:00.020)       0:08:41.980 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Monday 20 January 2025  06:20:16 -0500 (0:00:02.862)       0:08:44.843 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Monday 20 January 2025  06:20:16 -0500 (0:00:00.025)       0:08:44.868 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Monday 20 January 2025  06:20:16 -0500 (0:00:00.024)       0:08:44.893 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Monday 20 January 2025  06:20:16 -0500 (0:00:00.038)       0:08:44.931 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Monday 20 January 2025  06:20:16 -0500 (0:00:00.024)       0:08:44.955 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Monday 20 January 2025  06:20:16 -0500 (0:00:00.029)       0:08:44.984 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Monday 20 January 2025  06:20:16 -0500 (0:00:00.026)       0:08:45.011 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Monday 20 January 2025  06:20:16 -0500 (0:00:00.020)       0:08:45.032 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Monday 20 January 2025  06:20:16 -0500 (0:00:00.020)       0:08:45.053 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Monday 20 January 2025  06:20:16 -0500 (0:00:00.026)       0:08:45.079 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Monday 20 January 2025  06:20:16 -0500 (0:00:00.024)       0:08:45.104 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Monday 20 January 2025  06:20:16 -0500 (0:00:00.019)       0:08:45.124 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Monday 20 January 2025  06:20:16 -0500 (0:00:00.021)       0:08:45.146 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Monday 20 January 2025  06:20:16 -0500 (0:00:00.025)       0:08:45.171 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Monday 20 January 2025  06:20:16 -0500 (0:00:00.029)       0:08:45.201 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Monday 20 January 2025  06:20:16 -0500 (0:00:00.023)       0:08:45.224 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Monday 20 January 2025  06:20:16 -0500 (0:00:00.038)       0:08:45.263 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Monday 20 January 2025  06:20:16 -0500 (0:00:00.031)       0:08:45.294 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Monday 20 January 2025  06:20:16 -0500 (0:00:00.033)       0:08:45.327 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Monday 20 January 2025  06:20:16 -0500 (0:00:00.040)       0:08:45.368 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Monday 20 January 2025  06:20:16 -0500 (0:00:00.034)       0:08:45.402 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Monday 20 January 2025  06:20:16 -0500 (0:00:00.053)       0:08:45.456 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Monday 20 January 2025  06:20:16 -0500 (0:00:00.065)       0:08:45.522 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Monday 20 January 2025  06:20:16 -0500 (0:00:00.055)       0:08:45.578 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Monday 20 January 2025  06:20:16 -0500 (0:00:00.045)       0:08:45.623 ******** 
ok: [managed-node1] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Monday 20 January 2025  06:20:17 -0500 (0:00:00.442)       0:08:46.065 ******** 
ok: [managed-node1] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Monday 20 January 2025  06:20:17 -0500 (0:00:00.404)       0:08:46.470 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_expected_size": "4294967296"
    },
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Monday 20 January 2025  06:20:17 -0500 (0:00:00.038)       0:08:46.508 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Monday 20 January 2025  06:20:17 -0500 (0:00:00.036)       0:08:46.545 ******** 
ok: [managed-node1] => {
    "bytes": 10726680821,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Monday 20 January 2025  06:20:18 -0500 (0:00:00.446)       0:08:46.992 ******** 
skipping: [managed-node1] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Monday 20 January 2025  06:20:18 -0500 (0:00:00.039)       0:08:47.031 ******** 
skipping: [managed-node1] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Monday 20 January 2025  06:20:18 -0500 (0:00:00.041)       0:08:47.073 ******** 
skipping: [managed-node1] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Monday 20 January 2025  06:20:18 -0500 (0:00:00.029)       0:08:47.103 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Monday 20 January 2025  06:20:18 -0500 (0:00:00.032)       0:08:47.135 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Monday 20 January 2025  06:20:18 -0500 (0:00:00.025)       0:08:47.161 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Monday 20 January 2025  06:20:18 -0500 (0:00:00.026)       0:08:47.187 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Monday 20 January 2025  06:20:18 -0500 (0:00:00.025)       0:08:47.213 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Monday 20 January 2025  06:20:18 -0500 (0:00:00.020)       0:08:47.233 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Monday 20 January 2025  06:20:18 -0500 (0:00:00.032)       0:08:47.265 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Monday 20 January 2025  06:20:18 -0500 (0:00:00.021)       0:08:47.287 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Monday 20 January 2025  06:20:18 -0500 (0:00:00.020)       0:08:47.308 ******** 
skipping: [managed-node1] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Monday 20 January 2025  06:20:18 -0500 (0:00:00.060)       0:08:47.369 ******** 
skipping: [managed-node1] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Monday 20 January 2025  06:20:18 -0500 (0:00:00.022)       0:08:47.391 ******** 
skipping: [managed-node1] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Monday 20 January 2025  06:20:18 -0500 (0:00:00.020)       0:08:47.412 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Monday 20 January 2025  06:20:18 -0500 (0:00:00.020)       0:08:47.432 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Monday 20 January 2025  06:20:18 -0500 (0:00:00.020)       0:08:47.452 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Monday 20 January 2025  06:20:18 -0500 (0:00:00.020)       0:08:47.473 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Monday 20 January 2025  06:20:18 -0500 (0:00:00.023)       0:08:47.497 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Monday 20 January 2025  06:20:18 -0500 (0:00:00.025)       0:08:47.522 ******** 
ok: [managed-node1] => {
    "storage_test_actual_size": {
        "bytes": 4294967296,
        "changed": false,
        "failed": false,
        "lvm": "4g",
        "parted": "4GiB",
        "size": "4 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Monday 20 January 2025  06:20:18 -0500 (0:00:00.029)       0:08:47.552 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Monday 20 January 2025  06:20:18 -0500 (0:00:00.032)       0:08:47.585 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Monday 20 January 2025  06:20:18 -0500 (0:00:00.044)       0:08:47.630 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.028048",
    "end": "2025-01-20 06:20:19.199397",
    "rc": 0,
    "start": "2025-01-20 06:20:19.171349"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Monday 20 January 2025  06:20:19 -0500 (0:00:00.399)       0:08:48.029 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Monday 20 January 2025  06:20:19 -0500 (0:00:00.040)       0:08:48.069 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Monday 20 January 2025  06:20:19 -0500 (0:00:00.042)       0:08:48.111 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Monday 20 January 2025  06:20:19 -0500 (0:00:00.033)       0:08:48.145 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Monday 20 January 2025  06:20:19 -0500 (0:00:00.027)       0:08:48.173 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Monday 20 January 2025  06:20:19 -0500 (0:00:00.023)       0:08:48.196 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Monday 20 January 2025  06:20:19 -0500 (0:00:00.031)       0:08:48.227 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Monday 20 January 2025  06:20:19 -0500 (0:00:00.020)       0:08:48.248 ******** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Monday 20 January 2025  06:20:19 -0500 (0:00:00.024)       0:08:48.273 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Create a file] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/create-test-file.yml:12
Monday 20 January 2025  06:20:19 -0500 (0:00:00.041)       0:08:48.314 ******** 
changed: [managed-node1] => {
    "changed": true,
    "dest": "/opt/test1/quux",
    "gid": 0,
    "group": "root",
    "mode": "0644",
    "owner": "root",
    "secontext": "unconfined_u:object_r:unlabeled_t:s0",
    "size": 0,
    "state": "file",
    "uid": 0
}

TASK [Test for correct handling of safe_mode] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:483
Monday 20 January 2025  06:20:19 -0500 (0:00:00.368)       0:08:48.683 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node1

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Monday 20 January 2025  06:20:19 -0500 (0:00:00.042)       0:08:48.726 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_pools_global": [],
        "storage_safe_mode_global": true,
        "storage_volumes_global": []
    },
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Monday 20 January 2025  06:20:19 -0500 (0:00:00.028)       0:08:48.754 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:20:20 -0500 (0:00:00.029)       0:08:48.784 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:20:20 -0500 (0:00:00.034)       0:08:48.818 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:20:20 -0500 (0:00:00.044)       0:08:48.863 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:20:20 -0500 (0:00:00.073)       0:08:48.936 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:20:20 -0500 (0:00:00.022)       0:08:48.958 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:20:20 -0500 (0:00:00.026)       0:08:48.985 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:20:20 -0500 (0:00:00.021)       0:08:49.006 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:20:20 -0500 (0:00:00.024)       0:08:49.031 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:20:20 -0500 (0:00:00.057)       0:08:49.088 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:20:23 -0500 (0:00:02.859)       0:08:51.948 ******** 
ok: [managed-node1] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "encryption": true,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "yabbadabbadoo",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:20:23 -0500 (0:00:00.027)       0:08:51.976 ******** 
ok: [managed-node1] => {
    "storage_volumes": []
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:20:23 -0500 (0:00:00.026)       0:08:52.003 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "cryptsetup",
        "lvm2"
    ],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:20:27 -0500 (0:00:04.208)       0:08:56.211 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:20:27 -0500 (0:00:00.071)       0:08:56.283 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:20:27 -0500 (0:00:00.029)       0:08:56.313 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:20:27 -0500 (0:00:00.036)       0:08:56.350 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:20:27 -0500 (0:00:00.031)       0:08:56.382 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:20:30 -0500 (0:00:03.106)       0:08:59.488 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service": {
                "name": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
                "source": "systemd",
                "state": "inactive",
                "status": "generated"
            },
            "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service": {
                "name": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:20:32 -0500 (0:00:01.780)       0:09:01.269 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": [
            "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
            "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service"
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:20:32 -0500 (0:00:00.038)       0:09:01.307 ******** 
changed: [managed-node1] => (item=systemd-cryptsetup@luks\x2d940b35f5\x2d2180\x2d4a53\x2d8562\x2de5b2ccb3efe6.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "name": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "After": "dev-mapper-foo\\x2dtest1.device systemd-journald.socket cryptsetup-pre.target system-systemd\\x2dcryptsetup.slice",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "Before": "umount.target cryptsetup.target",
        "BindsTo": "dev-mapper-foo\\x2dtest1.device",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "yes",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "Conflicts": "umount.target",
        "ControlPID": "0",
        "DefaultDependencies": "no",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "Cryptography Setup for luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6",
        "DevicePolicy": "auto",
        "Documentation": "man:crypttab(5) man:systemd-cryptsetup-generator(8) man:systemd-cryptsetup@.service(8)",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6 /dev/mapper/foo-test1 -  ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
        "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-940b35f5-2180-4a53-8562-e5b2ccb3efe6 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "IgnoreOnIsolate": "yes",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "shared",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "0",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "65536",
        "LimitMEMLOCKSoft": "65536",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "262144",
        "LimitNOFILESoft": "1024",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadState": "loaded",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "NeedDaemonReload": "yes",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "yes",
        "RemoveIPC": "no",
        "RequiredBy": "cryptsetup.target",
        "Requires": "system-systemd\\x2dcryptsetup.slice",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "Slice": "system-systemd\\x2dcryptsetup.slice",
        "SourcePath": "/etc/crypttab",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "journal",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestamp": "Mon 2025-01-20 06:18:55 EST",
        "StateChangeTimestampMonotonic": "2753199841",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "infinity",
        "TimeoutStopUSec": "infinity",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "Type": "oneshot",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFilePreset": "disabled",
        "UnitFileState": "generated",
        "UtmpMode": "init",
        "WantedBy": "dev-mapper-foo\\x2dtest1.device",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}
changed: [managed-node1] => (item=systemd-cryptsetup@luk...d2180\x2d4a53\x2d8562\x2de5b2ccb3efe6.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "name": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.NoSuchUnit \"Unit systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service not found.\"",
        "LoadState": "not-found",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:20:33 -0500 (0:00:01.334)       0:09:02.642 ******** 
fatal: [managed-node1]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

cannot remove existing formatting on device 'test1' in safe mode due to adding encryption

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Monday 20 January 2025  06:20:38 -0500 (0:00:04.169)       0:09:06.811 ******** 
fatal: [managed-node1]: FAILED! => {
    "changed": false
}

MSG:

{'msg': "cannot remove existing formatting on device 'test1' in safe mode due to adding encryption", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'encryption_clevis_pin': None, 'encryption_tang_url': None, 'encryption_tang_thumbprint': None, 'grow_to_fill': False, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'shared': False, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': True, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': 'luks2', 'encryption_password': 'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}]}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'grow_to_fill': False, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, 'shared': False}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:20:38 -0500 (0:00:00.075)       0:09:06.886 ******** 
changed: [managed-node1] => (item=systemd-cryptsetup@luks\x2d940b35f5\x2d2180\x2d4a53\x2d8562\x2de5b2ccb3efe6.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "name": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/dev/null",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service is masked.\"",
        "LoadState": "masked",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luks\\x2d940b35f5\\x2d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFileState": "masked",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}
changed: [managed-node1] => (item=systemd-cryptsetup@luk...d2180\x2d4a53\x2d8562\x2de5b2ccb3efe6.service) => {
    "ansible_loop_var": "item",
    "changed": true,
    "item": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "name": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
    "status": {
        "ActiveEnterTimestampMonotonic": "0",
        "ActiveExitTimestampMonotonic": "0",
        "ActiveState": "inactive",
        "AllowIsolate": "no",
        "AllowedCPUs": "",
        "AllowedMemoryNodes": "",
        "AmbientCapabilities": "",
        "AssertResult": "no",
        "AssertTimestampMonotonic": "0",
        "BlockIOAccounting": "no",
        "BlockIOWeight": "[not set]",
        "CPUAccounting": "no",
        "CPUAffinity": "",
        "CPUAffinityFromNUMA": "no",
        "CPUQuotaPerSecUSec": "infinity",
        "CPUQuotaPeriodUSec": "infinity",
        "CPUSchedulingPolicy": "0",
        "CPUSchedulingPriority": "0",
        "CPUSchedulingResetOnFork": "no",
        "CPUShares": "[not set]",
        "CPUUsageNSec": "[not set]",
        "CPUWeight": "[not set]",
        "CacheDirectoryMode": "0755",
        "CanFreeze": "yes",
        "CanIsolate": "no",
        "CanReload": "no",
        "CanStart": "no",
        "CanStop": "yes",
        "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf",
        "CollectMode": "inactive",
        "ConditionResult": "no",
        "ConditionTimestampMonotonic": "0",
        "ConfigurationDirectoryMode": "0755",
        "ControlPID": "0",
        "DefaultDependencies": "yes",
        "DefaultMemoryLow": "0",
        "DefaultMemoryMin": "0",
        "Delegate": "no",
        "Description": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "DevicePolicy": "auto",
        "DynamicUser": "no",
        "EffectiveCPUs": "",
        "EffectiveMemoryNodes": "",
        "ExecMainCode": "0",
        "ExecMainExitTimestampMonotonic": "0",
        "ExecMainPID": "0",
        "ExecMainStartTimestampMonotonic": "0",
        "ExecMainStatus": "0",
        "FailureAction": "none",
        "FileDescriptorStoreMax": "0",
        "FragmentPath": "/dev/null",
        "FreezerState": "running",
        "GID": "[not set]",
        "GuessMainPID": "yes",
        "IOAccounting": "no",
        "IOSchedulingClass": "0",
        "IOSchedulingPriority": "0",
        "IOWeight": "[not set]",
        "IPAccounting": "no",
        "IPEgressBytes": "18446744073709551615",
        "IPEgressPackets": "18446744073709551615",
        "IPIngressBytes": "18446744073709551615",
        "IPIngressPackets": "18446744073709551615",
        "Id": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "IgnoreOnIsolate": "no",
        "IgnoreSIGPIPE": "yes",
        "InactiveEnterTimestampMonotonic": "0",
        "InactiveExitTimestampMonotonic": "0",
        "JobRunningTimeoutUSec": "infinity",
        "JobTimeoutAction": "none",
        "JobTimeoutUSec": "infinity",
        "KeyringMode": "private",
        "KillMode": "control-group",
        "KillSignal": "15",
        "LimitAS": "infinity",
        "LimitASSoft": "infinity",
        "LimitCORE": "infinity",
        "LimitCORESoft": "infinity",
        "LimitCPU": "infinity",
        "LimitCPUSoft": "infinity",
        "LimitDATA": "infinity",
        "LimitDATASoft": "infinity",
        "LimitFSIZE": "infinity",
        "LimitFSIZESoft": "infinity",
        "LimitLOCKS": "infinity",
        "LimitLOCKSSoft": "infinity",
        "LimitMEMLOCK": "67108864",
        "LimitMEMLOCKSoft": "67108864",
        "LimitMSGQUEUE": "819200",
        "LimitMSGQUEUESoft": "819200",
        "LimitNICE": "0",
        "LimitNICESoft": "0",
        "LimitNOFILE": "1048576",
        "LimitNOFILESoft": "1048576",
        "LimitNPROC": "14003",
        "LimitNPROCSoft": "14003",
        "LimitRSS": "infinity",
        "LimitRSSSoft": "infinity",
        "LimitRTPRIO": "0",
        "LimitRTPRIOSoft": "0",
        "LimitRTTIME": "infinity",
        "LimitRTTIMESoft": "infinity",
        "LimitSIGPENDING": "14003",
        "LimitSIGPENDINGSoft": "14003",
        "LimitSTACK": "infinity",
        "LimitSTACKSoft": "8388608",
        "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service is masked.\"",
        "LoadState": "masked",
        "LockPersonality": "no",
        "LogLevelMax": "-1",
        "LogRateLimitBurst": "0",
        "LogRateLimitIntervalUSec": "0",
        "LogsDirectoryMode": "0755",
        "MainPID": "0",
        "MemoryAccounting": "yes",
        "MemoryCurrent": "[not set]",
        "MemoryDenyWriteExecute": "no",
        "MemoryHigh": "infinity",
        "MemoryLimit": "infinity",
        "MemoryLow": "0",
        "MemoryMax": "infinity",
        "MemoryMin": "0",
        "MemorySwapMax": "infinity",
        "MountAPIVFS": "no",
        "MountFlags": "",
        "NFileDescriptorStore": "0",
        "NRestarts": "0",
        "NUMAMask": "",
        "NUMAPolicy": "n/a",
        "Names": "systemd-cryptsetup@luk...d2180\\x2d4a53\\x2d8562\\x2de5b2ccb3efe6.service",
        "NeedDaemonReload": "no",
        "Nice": "0",
        "NoNewPrivileges": "no",
        "NonBlocking": "no",
        "NotifyAccess": "none",
        "OOMScoreAdjust": "0",
        "OnFailureJobMode": "replace",
        "PermissionsStartOnly": "no",
        "Perpetual": "no",
        "PrivateDevices": "no",
        "PrivateMounts": "no",
        "PrivateNetwork": "no",
        "PrivateTmp": "no",
        "PrivateUsers": "no",
        "ProtectControlGroups": "no",
        "ProtectHome": "no",
        "ProtectKernelModules": "no",
        "ProtectKernelTunables": "no",
        "ProtectSystem": "no",
        "RefuseManualStart": "no",
        "RefuseManualStop": "no",
        "RemainAfterExit": "no",
        "RemoveIPC": "no",
        "Restart": "no",
        "RestartUSec": "100ms",
        "RestrictNamespaces": "no",
        "RestrictRealtime": "no",
        "RestrictSUIDSGID": "no",
        "Result": "success",
        "RootDirectoryStartOnly": "no",
        "RuntimeDirectoryMode": "0755",
        "RuntimeDirectoryPreserve": "no",
        "RuntimeMaxUSec": "infinity",
        "SameProcessGroup": "no",
        "SecureBits": "0",
        "SendSIGHUP": "no",
        "SendSIGKILL": "yes",
        "StandardError": "inherit",
        "StandardInput": "null",
        "StandardInputData": "",
        "StandardOutput": "inherit",
        "StartLimitAction": "none",
        "StartLimitBurst": "5",
        "StartLimitIntervalUSec": "10s",
        "StartupBlockIOWeight": "[not set]",
        "StartupCPUShares": "[not set]",
        "StartupCPUWeight": "[not set]",
        "StartupIOWeight": "[not set]",
        "StateChangeTimestampMonotonic": "0",
        "StateDirectoryMode": "0755",
        "StatusErrno": "0",
        "StopWhenUnneeded": "no",
        "SubState": "dead",
        "SuccessAction": "none",
        "SyslogFacility": "3",
        "SyslogLevel": "6",
        "SyslogLevelPrefix": "yes",
        "SyslogPriority": "30",
        "SystemCallErrorNumber": "0",
        "TTYReset": "no",
        "TTYVHangup": "no",
        "TTYVTDisallocate": "no",
        "TasksAccounting": "yes",
        "TasksCurrent": "[not set]",
        "TasksMax": "22405",
        "TimeoutStartUSec": "1min 30s",
        "TimeoutStopUSec": "1min 30s",
        "TimerSlackNSec": "50000",
        "Transient": "no",
        "UID": "[not set]",
        "UMask": "0022",
        "UnitFileState": "masked",
        "UtmpMode": "init",
        "WatchdogTimestampMonotonic": "0",
        "WatchdogUSec": "0"
    }
}

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Monday 20 January 2025  06:20:39 -0500 (0:00:01.833)       0:09:08.720 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Monday 20 January 2025  06:20:39 -0500 (0:00:00.036)       0:09:08.756 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Monday 20 January 2025  06:20:40 -0500 (0:00:00.045)       0:09:08.802 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Stat the file] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:11
Monday 20 January 2025  06:20:40 -0500 (0:00:00.030)       0:09:08.833 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737372019.8614757,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1737372019.8614757,
        "dev": 64768,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 131,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0644",
        "mtime": 1737372019.8614757,
        "nlink": 1,
        "path": "/opt/test1/quux",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "804596755",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Assert file presence] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-data-preservation.yml:16
Monday 20 January 2025  06:20:40 -0500 (0:00:00.418)       0:09:09.252 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Add encryption to the volume] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:507
Monday 20 January 2025  06:20:40 -0500 (0:00:00.025)       0:09:09.277 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:20:40 -0500 (0:00:00.063)       0:09:09.341 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:20:40 -0500 (0:00:00.031)       0:09:09.372 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:20:40 -0500 (0:00:00.025)       0:09:09.398 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:20:40 -0500 (0:00:00.056)       0:09:09.454 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:20:40 -0500 (0:00:00.021)       0:09:09.475 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:20:40 -0500 (0:00:00.022)       0:09:09.498 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:20:40 -0500 (0:00:00.022)       0:09:09.520 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:20:40 -0500 (0:00:00.020)       0:09:09.540 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:20:40 -0500 (0:00:00.047)       0:09:09.588 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:20:43 -0500 (0:00:02.872)       0:09:12.461 ******** 
ok: [managed-node1] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "encryption": true,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "yabbadabbadoo",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:20:43 -0500 (0:00:00.026)       0:09:12.487 ******** 
ok: [managed-node1] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:20:43 -0500 (0:00:00.022)       0:09:12.510 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "cryptsetup",
        "lvm2"
    ],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:20:47 -0500 (0:00:03.764)       0:09:16.274 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:20:47 -0500 (0:00:00.050)       0:09:16.325 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:20:47 -0500 (0:00:00.036)       0:09:16.361 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:20:47 -0500 (0:00:00.038)       0:09:16.400 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:20:47 -0500 (0:00:00.030)       0:09:16.430 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:20:50 -0500 (0:00:02.921)       0:09:19.352 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:20:52 -0500 (0:00:01.718)       0:09:21.070 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:20:52 -0500 (0:00:00.050)       0:09:21.121 ******** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:20:52 -0500 (0:00:00.029)       0:09:21.150 ******** 
changed: [managed-node1] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "xfs"
        },
        {
            "action": "create format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "luks"
        },
        {
            "action": "create device",
            "device": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
            "fs_type": "xfs"
        }
    ],
    "changed": true,
    "crypts": [
        {
            "backing_device": "/dev/mapper/foo-test1",
            "name": "luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
            "password": "-",
            "state": "present"
        }
    ],
    "leaves": [
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1",
        "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a"
    ],
    "mounts": [
        {
            "fstype": "xfs",
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "absent"
        },
        {
            "dump": 0,
            "fstype": "xfs",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "lvm2",
        "cryptsetup"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
                    "_kernel_device": "/dev/dm-1",
                    "_mount_id": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": true,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "xfs",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Monday 20 January 2025  06:21:07 -0500 (0:00:14.654)       0:09:35.805 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Monday 20 January 2025  06:21:07 -0500 (0:00:00.021)       0:09:35.826 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737372004.257438,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "a1522684f5b6a445a50f2611a4e0757a4aec1cf1",
        "ctime": 1737372004.254438,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 207618244,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1737372004.254438,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1393,
        "uid": 0,
        "version": "261612921",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Monday 20 January 2025  06:21:07 -0500 (0:00:00.365)       0:09:36.192 ******** 
ok: [managed-node1] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:21:07 -0500 (0:00:00.363)       0:09:36.555 ******** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Monday 20 January 2025  06:21:07 -0500 (0:00:00.019)       0:09:36.574 ******** 
ok: [managed-node1] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "xfs"
            },
            {
                "action": "create format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "luks"
            },
            {
                "action": "create device",
                "device": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
                "fs_type": "xfs"
            }
        ],
        "changed": true,
        "crypts": [
            {
                "backing_device": "/dev/mapper/foo-test1",
                "name": "luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
                "password": "-",
                "state": "present"
            }
        ],
        "failed": false,
        "leaves": [
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1",
            "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a"
        ],
        "mounts": [
            {
                "fstype": "xfs",
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "absent"
            },
            {
                "dump": 0,
                "fstype": "xfs",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "lvm2",
            "cryptsetup"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
                        "_kernel_device": "/dev/dm-1",
                        "_mount_id": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": true,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": "luks2",
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "xfs",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Monday 20 January 2025  06:21:07 -0500 (0:00:00.026)       0:09:36.601 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
                        "_kernel_device": "/dev/dm-1",
                        "_mount_id": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": true,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": "luks2",
                        "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "xfs",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Monday 20 January 2025  06:21:07 -0500 (0:00:00.026)       0:09:36.627 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Monday 20 January 2025  06:21:07 -0500 (0:00:00.023)       0:09:36.651 ******** 
changed: [managed-node1] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "fstype": "xfs",
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Monday 20 January 2025  06:21:08 -0500 (0:00:00.371)       0:09:37.023 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Monday 20 January 2025  06:21:08 -0500 (0:00:00.625)       0:09:37.649 ******** 
changed: [managed-node1] => (item={'src': '/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Monday 20 January 2025  06:21:09 -0500 (0:00:00.401)       0:09:38.050 ******** 
skipping: [managed-node1] => (item={'src': '/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "xfs",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Monday 20 January 2025  06:21:09 -0500 (0:00:00.041)       0:09:38.092 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Monday 20 January 2025  06:21:10 -0500 (0:00:00.692)       0:09:38.785 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737372008.1334474,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1737372006.1484425,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 220201157,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1737372006.1464424,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "3988354207",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Monday 20 January 2025  06:21:10 -0500 (0:00:00.446)       0:09:39.232 ******** 
changed: [managed-node1] => (item={'backing_device': '/dev/mapper/foo-test1', 'name': 'luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a', 'password': '-', 'state': 'present'}) => {
    "ansible_loop_var": "entry",
    "backup": "",
    "changed": true,
    "entry": {
        "backing_device": "/dev/mapper/foo-test1",
        "name": "luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
        "password": "-",
        "state": "present"
    }
}

MSG:

line added

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Monday 20 January 2025  06:21:10 -0500 (0:00:00.496)       0:09:39.728 ******** 
ok: [managed-node1]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:524
Monday 20 January 2025  06:21:11 -0500 (0:00:00.867)       0:09:40.596 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node1

TASK [Print out pool information] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Monday 20 January 2025  06:21:11 -0500 (0:00:00.054)       0:09:40.650 ******** 
ok: [managed-node1] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
                    "_kernel_device": "/dev/dm-1",
                    "_mount_id": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": true,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": "luks2",
                    "encryption_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "xfs",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Monday 20 January 2025  06:21:11 -0500 (0:00:00.034)       0:09:40.684 ******** 
skipping: [managed-node1] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Monday 20 January 2025  06:21:11 -0500 (0:00:00.025)       0:09:40.710 ******** 
ok: [managed-node1] => {
    "changed": false,
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "crypto_LUKS",
            "label": "",
            "mountpoint": "",
            "name": "/dev/mapper/foo-test1",
            "size": "4G",
            "type": "lvm",
            "uuid": "9466c2a2-5725-4c1f-879d-9d629fc1c71a"
        },
        "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
            "size": "4G",
            "type": "crypt",
            "uuid": "c2eb5628-68f2-4de4-bf77-16bd97604526"
        },
        "/dev/sda": {
            "fstype": "LVM2_member",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "kSV36d-iWVA-92sg-L6wf-6hE7-Eik8-GLOUAJ"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Monday 20 January 2025  06:21:12 -0500 (0:00:00.369)       0:09:41.080 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003108",
    "end": "2025-01-20 06:21:12.654248",
    "rc": 0,
    "start": "2025-01-20 06:21:12.651140"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a /opt/test1 xfs defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Monday 20 January 2025  06:21:12 -0500 (0:00:00.392)       0:09:41.473 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002556",
    "end": "2025-01-20 06:21:13.003593",
    "failed_when_result": false,
    "rc": 0,
    "start": "2025-01-20 06:21:13.001037"
}

STDOUT:

luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a /dev/mapper/foo-test1 -

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Monday 20 January 2025  06:21:13 -0500 (0:00:00.357)       0:09:41.830 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node1

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Monday 20 January 2025  06:21:13 -0500 (0:00:00.052)       0:09:41.883 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Monday 20 January 2025  06:21:13 -0500 (0:00:00.024)       0:09:41.908 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "vgs",
        "--noheadings",
        "--binary",
        "-o",
        "shared",
        "foo"
    ],
    "delta": "0:00:00.025746",
    "end": "2025-01-20 06:21:13.472495",
    "rc": 0,
    "start": "2025-01-20 06:21:13.446749"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Monday 20 January 2025  06:21:13 -0500 (0:00:00.385)       0:09:42.293 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Monday 20 January 2025  06:21:13 -0500 (0:00:00.030)       0:09:42.324 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node1

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Monday 20 January 2025  06:21:13 -0500 (0:00:00.047)       0:09:42.372 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Monday 20 January 2025  06:21:13 -0500 (0:00:00.032)       0:09:42.405 ******** 
ok: [managed-node1] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sda",
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Monday 20 January 2025  06:21:14 -0500 (0:00:00.376)       0:09:42.782 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Monday 20 January 2025  06:21:14 -0500 (0:00:00.028)       0:09:42.810 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Monday 20 January 2025  06:21:14 -0500 (0:00:00.036)       0:09:42.846 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Monday 20 January 2025  06:21:14 -0500 (0:00:00.038)       0:09:42.885 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Monday 20 January 2025  06:21:14 -0500 (0:00:00.034)       0:09:42.919 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Monday 20 January 2025  06:21:14 -0500 (0:00:00.033)       0:09:42.953 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Monday 20 January 2025  06:21:14 -0500 (0:00:00.022)       0:09:42.976 ******** 
ok: [managed-node1] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Monday 20 January 2025  06:21:14 -0500 (0:00:00.033)       0:09:43.010 ******** 
ok: [managed-node1] => {
    "changed": false,
    "failed_when_result": false,
    "rc": 1
}

STDERR:

Shared connection to 10.31.8.138 closed.



MSG:

non-zero return code

TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:74
Monday 20 January 2025  06:21:14 -0500 (0:00:00.410)       0:09:43.420 ******** 
skipping: [managed-node1] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:84
Monday 20 January 2025  06:21:14 -0500 (0:00:00.025)       0:09:43.446 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node1

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Monday 20 January 2025  06:21:14 -0500 (0:00:00.040)       0:09:43.486 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Monday 20 January 2025  06:21:14 -0500 (0:00:00.021)       0:09:43.508 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Monday 20 January 2025  06:21:14 -0500 (0:00:00.020)       0:09:43.529 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Monday 20 January 2025  06:21:14 -0500 (0:00:00.022)       0:09:43.551 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Monday 20 January 2025  06:21:14 -0500 (0:00:00.020)       0:09:43.571 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Monday 20 January 2025  06:21:14 -0500 (0:00:00.019)       0:09:43.591 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Monday 20 January 2025  06:21:14 -0500 (0:00:00.019)       0:09:43.610 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Monday 20 January 2025  06:21:14 -0500 (0:00:00.020)       0:09:43.631 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Monday 20 January 2025  06:21:14 -0500 (0:00:00.020)       0:09:43.651 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Monday 20 January 2025  06:21:14 -0500 (0:00:00.028)       0:09:43.680 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Monday 20 January 2025  06:21:14 -0500 (0:00:00.020)       0:09:43.700 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:87
Monday 20 January 2025  06:21:14 -0500 (0:00:00.030)       0:09:43.730 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node1

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Monday 20 January 2025  06:21:15 -0500 (0:00:00.066)       0:09:43.797 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node1

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Monday 20 January 2025  06:21:15 -0500 (0:00:00.071)       0:09:43.868 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Monday 20 January 2025  06:21:15 -0500 (0:00:00.034)       0:09:43.903 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Monday 20 January 2025  06:21:15 -0500 (0:00:00.034)       0:09:43.938 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Monday 20 January 2025  06:21:15 -0500 (0:00:00.029)       0:09:43.967 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Monday 20 January 2025  06:21:15 -0500 (0:00:00.032)       0:09:44.000 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Monday 20 January 2025  06:21:15 -0500 (0:00:00.044)       0:09:44.044 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Monday 20 January 2025  06:21:15 -0500 (0:00:00.057)       0:09:44.102 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:90
Monday 20 January 2025  06:21:15 -0500 (0:00:00.025)       0:09:44.128 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node1

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Monday 20 January 2025  06:21:15 -0500 (0:00:00.063)       0:09:44.192 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node1

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Monday 20 January 2025  06:21:15 -0500 (0:00:00.170)       0:09:44.362 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Monday 20 January 2025  06:21:15 -0500 (0:00:00.026)       0:09:44.389 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Monday 20 January 2025  06:21:15 -0500 (0:00:00.021)       0:09:44.411 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Monday 20 January 2025  06:21:15 -0500 (0:00:00.020)       0:09:44.432 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:93
Monday 20 January 2025  06:21:15 -0500 (0:00:00.021)       0:09:44.453 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node1

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Monday 20 January 2025  06:21:15 -0500 (0:00:00.046)       0:09:44.499 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Monday 20 January 2025  06:21:15 -0500 (0:00:00.025)       0:09:44.525 ******** 
skipping: [managed-node1] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Monday 20 January 2025  06:21:15 -0500 (0:00:00.044)       0:09:44.569 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node1

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Monday 20 January 2025  06:21:15 -0500 (0:00:00.068)       0:09:44.638 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Monday 20 January 2025  06:21:15 -0500 (0:00:00.036)       0:09:44.675 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Monday 20 January 2025  06:21:15 -0500 (0:00:00.037)       0:09:44.713 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Monday 20 January 2025  06:21:15 -0500 (0:00:00.044)       0:09:44.757 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Monday 20 January 2025  06:21:16 -0500 (0:00:00.048)       0:09:44.806 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Monday 20 January 2025  06:21:16 -0500 (0:00:00.030)       0:09:44.836 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Monday 20 January 2025  06:21:16 -0500 (0:00:00.033)       0:09:44.870 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:96
Monday 20 January 2025  06:21:16 -0500 (0:00:00.025)       0:09:44.896 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node1

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Monday 20 January 2025  06:21:16 -0500 (0:00:00.066)       0:09:44.962 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node1

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Monday 20 January 2025  06:21:16 -0500 (0:00:00.054)       0:09:45.017 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Monday 20 January 2025  06:21:16 -0500 (0:00:00.023)       0:09:45.040 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Monday 20 January 2025  06:21:16 -0500 (0:00:00.022)       0:09:45.062 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Monday 20 January 2025  06:21:16 -0500 (0:00:00.019)       0:09:45.082 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Monday 20 January 2025  06:21:16 -0500 (0:00:00.022)       0:09:45.104 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Monday 20 January 2025  06:21:16 -0500 (0:00:00.020)       0:09:45.124 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Monday 20 January 2025  06:21:16 -0500 (0:00:00.025)       0:09:45.150 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:99
Monday 20 January 2025  06:21:16 -0500 (0:00:00.019)       0:09:45.170 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node1

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Monday 20 January 2025  06:21:16 -0500 (0:00:00.079)       0:09:45.249 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Monday 20 January 2025  06:21:16 -0500 (0:00:00.023)       0:09:45.272 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Monday 20 January 2025  06:21:16 -0500 (0:00:00.022)       0:09:45.295 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Monday 20 January 2025  06:21:16 -0500 (0:00:00.028)       0:09:45.323 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Monday 20 January 2025  06:21:16 -0500 (0:00:00.029)       0:09:45.352 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Monday 20 January 2025  06:21:16 -0500 (0:00:00.026)       0:09:45.379 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:102
Monday 20 January 2025  06:21:16 -0500 (0:00:00.021)       0:09:45.401 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Monday 20 January 2025  06:21:16 -0500 (0:00:00.022)       0:09:45.423 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node1

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Monday 20 January 2025  06:21:16 -0500 (0:00:00.037)       0:09:45.461 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Monday 20 January 2025  06:21:16 -0500 (0:00:00.027)       0:09:45.488 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node1

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Monday 20 January 2025  06:21:16 -0500 (0:00:00.144)       0:09:45.633 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Monday 20 January 2025  06:21:16 -0500 (0:00:00.067)       0:09:45.700 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Monday 20 January 2025  06:21:16 -0500 (0:00:00.046)       0:09:45.747 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Monday 20 January 2025  06:21:17 -0500 (0:00:00.037)       0:09:45.784 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Monday 20 January 2025  06:21:17 -0500 (0:00:00.041)       0:09:45.826 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Monday 20 January 2025  06:21:17 -0500 (0:00:00.030)       0:09:45.857 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Monday 20 January 2025  06:21:17 -0500 (0:00:00.030)       0:09:45.888 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Monday 20 January 2025  06:21:17 -0500 (0:00:00.025)       0:09:45.913 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Monday 20 January 2025  06:21:17 -0500 (0:00:00.023)       0:09:45.936 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Monday 20 January 2025  06:21:17 -0500 (0:00:00.022)       0:09:45.959 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Monday 20 January 2025  06:21:17 -0500 (0:00:00.102)       0:09:46.061 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Monday 20 January 2025  06:21:17 -0500 (0:00:00.034)       0:09:46.096 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 xfs defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Monday 20 January 2025  06:21:17 -0500 (0:00:00.054)       0:09:46.151 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Monday 20 January 2025  06:21:17 -0500 (0:00:00.030)       0:09:46.181 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Monday 20 January 2025  06:21:17 -0500 (0:00:00.030)       0:09:46.212 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Monday 20 January 2025  06:21:17 -0500 (0:00:00.028)       0:09:46.240 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Monday 20 January 2025  06:21:17 -0500 (0:00:00.035)       0:09:46.275 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Monday 20 January 2025  06:21:17 -0500 (0:00:00.025)       0:09:46.301 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Monday 20 January 2025  06:21:17 -0500 (0:00:00.033)       0:09:46.334 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Monday 20 January 2025  06:21:17 -0500 (0:00:00.041)       0:09:46.375 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737372066.7805893,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1737372066.7805893,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 702121,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1737372066.7805893,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Monday 20 January 2025  06:21:18 -0500 (0:00:00.433)       0:09:46.809 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Monday 20 January 2025  06:21:18 -0500 (0:00:00.043)       0:09:46.852 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Monday 20 January 2025  06:21:18 -0500 (0:00:00.032)       0:09:46.885 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Monday 20 January 2025  06:21:18 -0500 (0:00:00.031)       0:09:46.916 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Monday 20 January 2025  06:21:18 -0500 (0:00:00.036)       0:09:46.953 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Monday 20 January 2025  06:21:18 -0500 (0:00:00.024)       0:09:46.977 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Monday 20 January 2025  06:21:18 -0500 (0:00:00.036)       0:09:47.014 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737372066.9255898,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1737372066.9255898,
        "dev": 6,
        "device_type": 64769,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 719681,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1737372066.9255898,
        "nlink": 1,
        "path": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Monday 20 January 2025  06:21:18 -0500 (0:00:00.406)       0:09:47.420 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Monday 20 January 2025  06:21:21 -0500 (0:00:02.990)       0:09:50.411 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cryptsetup",
        "luksDump",
        "/dev/mapper/foo-test1"
    ],
    "delta": "0:00:00.010339",
    "end": "2025-01-20 06:21:22.065069",
    "rc": 0,
    "start": "2025-01-20 06:21:22.054730"
}

STDOUT:

LUKS header information
Version:       	2
Epoch:         	3
Metadata area: 	16384 [bytes]
Keyslots area: 	16744448 [bytes]
UUID:          	9466c2a2-5725-4c1f-879d-9d629fc1c71a
Label:         	(no label)
Subsystem:     	(no subsystem)
Flags:       	(no flags)

Data segments:
  0: crypt
	offset: 16777216 [bytes]
	length: (whole device)
	cipher: aes-xts-plain64
	sector: 512 [bytes]

Keyslots:
  0: luks2
	Key:        512 bits
	Priority:   normal
	Cipher:     aes-xts-plain64
	Cipher key: 512 bits
	PBKDF:      argon2i
	Time cost:  4
	Memory:     786171
	Threads:    2
	Salt:       98 a6 32 86 09 d0 c9 0c 0d 4c 9b d4 3b d2 53 e4 
	            84 aa d4 e9 c2 7b 1a d0 79 8b 1a 07 2e 0f 65 12 
	AF stripes: 4000
	AF hash:    sha256
	Area offset:32768 [bytes]
	Area length:258048 [bytes]
	Digest ID:  0
Tokens:
Digests:
  0: pbkdf2
	Hash:       sha256
	Iterations: 120029
	Salt:       d0 ba 4d 61 72 70 7c 80 ee 3e 86 8b 0d 15 f8 41 
	            0a 03 1d ac ab 96 29 fa 1b a7 4f 6a e2 9d 90 6b 
	Digest:     af 6d fb d0 d2 f4 1b 45 18 57 ed f0 c9 16 d0 8a 
	            61 65 17 ea 68 d0 64 75 a6 f3 0b 33 35 cf a0 20 

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Monday 20 January 2025  06:21:22 -0500 (0:00:00.542)       0:09:50.953 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Monday 20 January 2025  06:21:22 -0500 (0:00:00.074)       0:09:51.028 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Monday 20 January 2025  06:21:22 -0500 (0:00:00.059)       0:09:51.087 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Monday 20 January 2025  06:21:22 -0500 (0:00:00.079)       0:09:51.167 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Monday 20 January 2025  06:21:22 -0500 (0:00:00.082)       0:09:51.250 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Monday 20 January 2025  06:21:22 -0500 (0:00:00.050)       0:09:51.300 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Monday 20 January 2025  06:21:22 -0500 (0:00:00.041)       0:09:51.342 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Monday 20 January 2025  06:21:22 -0500 (0:00:00.040)       0:09:51.382 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [
            "luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a /dev/mapper/foo-test1 -"
        ],
        "_storage_test_expected_crypttab_entries": "1",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Monday 20 January 2025  06:21:22 -0500 (0:00:00.046)       0:09:51.429 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Monday 20 January 2025  06:21:22 -0500 (0:00:00.044)       0:09:51.474 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Monday 20 January 2025  06:21:22 -0500 (0:00:00.051)       0:09:51.526 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Monday 20 January 2025  06:21:22 -0500 (0:00:00.044)       0:09:51.570 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Monday 20 January 2025  06:21:22 -0500 (0:00:00.053)       0:09:51.623 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Monday 20 January 2025  06:21:22 -0500 (0:00:00.042)       0:09:51.666 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Monday 20 January 2025  06:21:22 -0500 (0:00:00.041)       0:09:51.707 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Monday 20 January 2025  06:21:22 -0500 (0:00:00.032)       0:09:51.740 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Monday 20 January 2025  06:21:23 -0500 (0:00:00.030)       0:09:51.770 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Monday 20 January 2025  06:21:23 -0500 (0:00:00.030)       0:09:51.801 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Monday 20 January 2025  06:21:23 -0500 (0:00:00.034)       0:09:51.835 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Monday 20 January 2025  06:21:23 -0500 (0:00:00.032)       0:09:51.867 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Monday 20 January 2025  06:21:23 -0500 (0:00:00.030)       0:09:51.898 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Monday 20 January 2025  06:21:23 -0500 (0:00:00.031)       0:09:51.929 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Monday 20 January 2025  06:21:23 -0500 (0:00:00.031)       0:09:51.961 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Monday 20 January 2025  06:21:23 -0500 (0:00:00.030)       0:09:51.992 ******** 
ok: [managed-node1] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Monday 20 January 2025  06:21:23 -0500 (0:00:00.413)       0:09:52.405 ******** 
ok: [managed-node1] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Monday 20 January 2025  06:21:24 -0500 (0:00:00.408)       0:09:52.814 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_expected_size": "4294967296"
    },
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Monday 20 January 2025  06:21:24 -0500 (0:00:00.049)       0:09:52.863 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Monday 20 January 2025  06:21:24 -0500 (0:00:00.031)       0:09:52.894 ******** 
ok: [managed-node1] => {
    "bytes": 10726680821,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Monday 20 January 2025  06:21:24 -0500 (0:00:00.404)       0:09:53.298 ******** 
skipping: [managed-node1] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Monday 20 January 2025  06:21:24 -0500 (0:00:00.043)       0:09:53.342 ******** 
skipping: [managed-node1] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Monday 20 January 2025  06:21:24 -0500 (0:00:00.046)       0:09:53.389 ******** 
skipping: [managed-node1] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Monday 20 January 2025  06:21:24 -0500 (0:00:00.041)       0:09:53.430 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Monday 20 January 2025  06:21:24 -0500 (0:00:00.039)       0:09:53.469 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Monday 20 January 2025  06:21:24 -0500 (0:00:00.031)       0:09:53.501 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Monday 20 January 2025  06:21:24 -0500 (0:00:00.030)       0:09:53.531 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Monday 20 January 2025  06:21:24 -0500 (0:00:00.031)       0:09:53.563 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Monday 20 January 2025  06:21:24 -0500 (0:00:00.033)       0:09:53.597 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Monday 20 January 2025  06:21:24 -0500 (0:00:00.030)       0:09:53.628 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Monday 20 January 2025  06:21:24 -0500 (0:00:00.031)       0:09:53.659 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Monday 20 January 2025  06:21:24 -0500 (0:00:00.033)       0:09:53.692 ******** 
skipping: [managed-node1] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Monday 20 January 2025  06:21:24 -0500 (0:00:00.031)       0:09:53.723 ******** 
skipping: [managed-node1] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Monday 20 January 2025  06:21:25 -0500 (0:00:00.044)       0:09:53.768 ******** 
skipping: [managed-node1] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Monday 20 January 2025  06:21:25 -0500 (0:00:00.034)       0:09:53.803 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Monday 20 January 2025  06:21:25 -0500 (0:00:00.033)       0:09:53.836 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Monday 20 January 2025  06:21:25 -0500 (0:00:00.032)       0:09:53.868 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Monday 20 January 2025  06:21:25 -0500 (0:00:00.034)       0:09:53.903 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Monday 20 January 2025  06:21:25 -0500 (0:00:00.031)       0:09:53.935 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Monday 20 January 2025  06:21:25 -0500 (0:00:00.034)       0:09:53.969 ******** 
ok: [managed-node1] => {
    "storage_test_actual_size": {
        "bytes": 4294967296,
        "changed": false,
        "failed": false,
        "lvm": "4g",
        "parted": "4GiB",
        "size": "4 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Monday 20 January 2025  06:21:25 -0500 (0:00:00.036)       0:09:54.006 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Monday 20 January 2025  06:21:25 -0500 (0:00:00.035)       0:09:54.042 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Monday 20 January 2025  06:21:25 -0500 (0:00:00.044)       0:09:54.087 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.024984",
    "end": "2025-01-20 06:21:25.686500",
    "rc": 0,
    "start": "2025-01-20 06:21:25.661516"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Monday 20 January 2025  06:21:25 -0500 (0:00:00.463)       0:09:54.550 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Monday 20 January 2025  06:21:25 -0500 (0:00:00.116)       0:09:54.666 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Monday 20 January 2025  06:21:25 -0500 (0:00:00.046)       0:09:54.713 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Monday 20 January 2025  06:21:25 -0500 (0:00:00.037)       0:09:54.750 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Monday 20 January 2025  06:21:26 -0500 (0:00:00.053)       0:09:54.804 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Monday 20 January 2025  06:21:26 -0500 (0:00:00.051)       0:09:54.856 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Monday 20 January 2025  06:21:26 -0500 (0:00:00.040)       0:09:54.896 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Monday 20 January 2025  06:21:26 -0500 (0:00:00.037)       0:09:54.934 ******** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Monday 20 January 2025  06:21:26 -0500 (0:00:00.032)       0:09:54.966 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Clean up] ****************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:527
Monday 20 January 2025  06:21:26 -0500 (0:00:00.036)       0:09:55.003 ******** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Monday 20 January 2025  06:21:26 -0500 (0:00:00.122)       0:09:55.125 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Monday 20 January 2025  06:21:26 -0500 (0:00:00.104)       0:09:55.229 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Monday 20 January 2025  06:21:26 -0500 (0:00:00.065)       0:09:55.295 ******** 
skipping: [managed-node1] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node1] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node1] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Monday 20 January 2025  06:21:26 -0500 (0:00:00.116)       0:09:55.412 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Monday 20 January 2025  06:21:26 -0500 (0:00:00.043)       0:09:55.455 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Monday 20 January 2025  06:21:26 -0500 (0:00:00.040)       0:09:55.496 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Monday 20 January 2025  06:21:26 -0500 (0:00:00.032)       0:09:55.528 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Monday 20 January 2025  06:21:26 -0500 (0:00:00.079)       0:09:55.607 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Monday 20 January 2025  06:21:26 -0500 (0:00:00.132)       0:09:55.740 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Monday 20 January 2025  06:21:29 -0500 (0:00:03.002)       0:09:58.743 ******** 
ok: [managed-node1] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Monday 20 January 2025  06:21:30 -0500 (0:00:00.023)       0:09:58.766 ******** 
ok: [managed-node1] => {
    "storage_volumes": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "state": "absent",
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Monday 20 January 2025  06:21:30 -0500 (0:00:00.028)       0:09:58.794 ******** 
ok: [managed-node1] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Monday 20 January 2025  06:21:34 -0500 (0:00:04.300)       0:10:03.095 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node1

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Monday 20 January 2025  06:21:34 -0500 (0:00:00.074)       0:10:03.169 ******** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Monday 20 January 2025  06:21:34 -0500 (0:00:00.050)       0:10:03.219 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Monday 20 January 2025  06:21:34 -0500 (0:00:00.033)       0:10:03.253 ******** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Monday 20 January 2025  06:21:34 -0500 (0:00:00.032)       0:10:03.285 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Monday 20 January 2025  06:21:37 -0500 (0:00:02.957)       0:10:06.243 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Monday 20 January 2025  06:21:39 -0500 (0:00:01.694)       0:10:07.937 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Monday 20 January 2025  06:21:39 -0500 (0:00:00.033)       0:10:07.971 ******** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Monday 20 January 2025  06:21:39 -0500 (0:00:00.019)       0:10:07.990 ******** 
changed: [managed-node1] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
            "fs_type": "xfs"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "luks"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "destroy device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/sda",
            "fs_type": "lvmpv"
        }
    ],
    "changed": true,
    "crypts": [
        {
            "backing_device": "/dev/mapper/foo-test1",
            "name": "luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
            "password": "-",
            "state": "absent"
        }
    ],
    "leaves": [
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1"
    ],
    "mounts": [
        {
            "fstype": "xfs",
            "path": "/opt/test1",
            "src": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
            "state": "absent"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [],
    "volumes": [
        {
            "_device": "/dev/sda",
            "_mount_id": "UUID=kSV36d-iWVA-92sg-L6wf-6hE7-Eik8-GLOUAJ",
            "_raw_device": "/dev/sda",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "fs_create_options": "",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "lvmpv",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_group": null,
            "mount_mode": null,
            "mount_options": "defaults",
            "mount_passno": 0,
            "mount_point": null,
            "mount_user": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "raid_stripe_size": null,
            "size": 10737418240,
            "state": "absent",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Monday 20 January 2025  06:21:43 -0500 (0:00:04.703)       0:10:12.694 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Monday 20 January 2025  06:21:43 -0500 (0:00:00.028)       0:10:12.722 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737372069.2005954,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "47e197eaf954e27eeb6de2aa57ad93a5e5b6a960",
        "ctime": 1737372069.1975954,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 207618244,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1737372069.1975954,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1425,
        "uid": 0,
        "version": "261612921",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Monday 20 January 2025  06:21:44 -0500 (0:00:00.371)       0:10:13.094 ******** 
ok: [managed-node1] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Monday 20 January 2025  06:21:44 -0500 (0:00:00.393)       0:10:13.487 ******** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Monday 20 January 2025  06:21:44 -0500 (0:00:00.020)       0:10:13.508 ******** 
ok: [managed-node1] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
                "fs_type": "xfs"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "luks"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "destroy device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/sda",
                "fs_type": "lvmpv"
            }
        ],
        "changed": true,
        "crypts": [
            {
                "backing_device": "/dev/mapper/foo-test1",
                "name": "luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
                "password": "-",
                "state": "absent"
            }
        ],
        "failed": false,
        "leaves": [
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1"
        ],
        "mounts": [
            {
                "fstype": "xfs",
                "path": "/opt/test1",
                "src": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
                "state": "absent"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [],
        "volumes": [
            {
                "_device": "/dev/sda",
                "_mount_id": "UUID=kSV36d-iWVA-92sg-L6wf-6hE7-Eik8-GLOUAJ",
                "_raw_device": "/dev/sda",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "fs_create_options": "",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "lvmpv",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_group": null,
                "mount_mode": null,
                "mount_options": "defaults",
                "mount_passno": 0,
                "mount_point": null,
                "mount_user": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "raid_stripe_size": null,
                "size": 10737418240,
                "state": "absent",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Monday 20 January 2025  06:21:44 -0500 (0:00:00.029)       0:10:13.537 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Monday 20 January 2025  06:21:44 -0500 (0:00:00.025)       0:10:13.563 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_volumes_list": [
            {
                "_device": "/dev/sda",
                "_mount_id": "UUID=kSV36d-iWVA-92sg-L6wf-6hE7-Eik8-GLOUAJ",
                "_raw_device": "/dev/sda",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "fs_create_options": "",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "lvmpv",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_group": null,
                "mount_mode": null,
                "mount_options": "defaults",
                "mount_passno": 0,
                "mount_point": null,
                "mount_user": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "raid_stripe_size": null,
                "size": 10737418240,
                "state": "absent",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Monday 20 January 2025  06:21:44 -0500 (0:00:00.031)       0:10:13.595 ******** 
changed: [managed-node1] => (item={'src': '/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "xfs",
    "mount_info": {
        "fstype": "xfs",
        "path": "/opt/test1",
        "src": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Monday 20 January 2025  06:21:45 -0500 (0:00:00.385)       0:10:13.980 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Monday 20 January 2025  06:21:45 -0500 (0:00:00.657)       0:10:14.638 ******** 

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Monday 20 January 2025  06:21:45 -0500 (0:00:00.032)       0:10:14.671 ******** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Monday 20 January 2025  06:21:45 -0500 (0:00:00.031)       0:10:14.702 ******** 
ok: [managed-node1] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Monday 20 January 2025  06:21:46 -0500 (0:00:00.663)       0:10:15.366 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737372073.0026045,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "a2c1476afe15a4b59048358e0add9e93af3618aa",
        "ctime": 1737372070.8875995,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 375390340,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0600",
        "mtime": 1737372070.8855994,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 66,
        "uid": 0,
        "version": "947035467",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Monday 20 January 2025  06:21:47 -0500 (0:00:00.407)       0:10:15.773 ******** 
changed: [managed-node1] => (item={'backing_device': '/dev/mapper/foo-test1', 'name': 'luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a', 'password': '-', 'state': 'absent'}) => {
    "ansible_loop_var": "entry",
    "backup": "",
    "changed": true,
    "entry": {
        "backing_device": "/dev/mapper/foo-test1",
        "name": "luks-9466c2a2-5725-4c1f-879d-9d629fc1c71a",
        "password": "-",
        "state": "absent"
    },
    "found": 1
}

MSG:

1 line(s) removed

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Monday 20 January 2025  06:21:47 -0500 (0:00:00.377)       0:10:16.151 ******** 
ok: [managed-node1]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/tests_luks2.yml:537
Monday 20 January 2025  06:21:48 -0500 (0:00:00.804)       0:10:16.955 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node1

TASK [Print out pool information] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Monday 20 January 2025  06:21:48 -0500 (0:00:00.062)       0:10:17.018 ******** 
skipping: [managed-node1] => {}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Monday 20 January 2025  06:21:48 -0500 (0:00:00.030)       0:10:17.049 ******** 
ok: [managed-node1] => {
    "_storage_volumes_list": [
        {
            "_device": "/dev/sda",
            "_mount_id": "UUID=kSV36d-iWVA-92sg-L6wf-6hE7-Eik8-GLOUAJ",
            "_raw_device": "/dev/sda",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "fs_create_options": "",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "lvmpv",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_group": null,
            "mount_mode": null,
            "mount_options": "defaults",
            "mount_passno": 0,
            "mount_point": null,
            "mount_user": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "raid_stripe_size": null,
            "size": 10737418240,
            "state": "absent",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Monday 20 January 2025  06:21:48 -0500 (0:00:00.037)       0:10:17.086 ******** 
ok: [managed-node1] => {
    "changed": false,
    "info": {
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Monday 20 January 2025  06:21:48 -0500 (0:00:00.479)       0:10:17.566 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.002643",
    "end": "2025-01-20 06:21:49.251881",
    "rc": 0,
    "start": "2025-01-20 06:21:49.249238"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Monday 20 January 2025  06:21:49 -0500 (0:00:00.519)       0:10:18.086 ******** 
ok: [managed-node1] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002797",
    "end": "2025-01-20 06:21:49.630559",
    "failed_when_result": false,
    "rc": 0,
    "start": "2025-01-20 06:21:49.627762"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Monday 20 January 2025  06:21:49 -0500 (0:00:00.383)       0:10:18.470 ******** 

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Monday 20 January 2025  06:21:49 -0500 (0:00:00.033)       0:10:18.504 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node1

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Monday 20 January 2025  06:21:49 -0500 (0:00:00.059)       0:10:18.564 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": false,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Monday 20 January 2025  06:21:49 -0500 (0:00:00.069)       0:10:18.633 ******** 
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node1
included: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node1

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Monday 20 January 2025  06:21:50 -0500 (0:00:00.266)       0:10:18.900 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/sda"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Monday 20 January 2025  06:21:50 -0500 (0:00:00.043)       0:10:18.944 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Monday 20 January 2025  06:21:50 -0500 (0:00:00.040)       0:10:18.984 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Monday 20 January 2025  06:21:50 -0500 (0:00:00.040)       0:10:19.025 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Monday 20 January 2025  06:21:50 -0500 (0:00:00.025)       0:10:19.051 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Monday 20 January 2025  06:21:50 -0500 (0:00:00.029)       0:10:19.080 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Monday 20 January 2025  06:21:50 -0500 (0:00:00.033)       0:10:19.114 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Monday 20 January 2025  06:21:50 -0500 (0:00:00.030)       0:10:19.144 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Monday 20 January 2025  06:21:50 -0500 (0:00:00.028)       0:10:19.172 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Monday 20 January 2025  06:21:50 -0500 (0:00:00.030)       0:10:19.202 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Monday 20 January 2025  06:21:50 -0500 (0:00:00.030)       0:10:19.233 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Monday 20 January 2025  06:21:50 -0500 (0:00:00.028)       0:10:19.262 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "0",
        "storage_test_fstab_expected_mount_options_matches": "0",
        "storage_test_fstab_expected_mount_point_matches": "0",
        "storage_test_fstab_id_matches": [],
        "storage_test_fstab_mount_options_matches": [],
        "storage_test_fstab_mount_point_matches": []
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Monday 20 January 2025  06:21:50 -0500 (0:00:00.062)       0:10:19.324 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Monday 20 January 2025  06:21:50 -0500 (0:00:00.029)       0:10:19.354 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Monday 20 January 2025  06:21:50 -0500 (0:00:00.036)       0:10:19.391 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Monday 20 January 2025  06:21:50 -0500 (0:00:00.029)       0:10:19.420 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Monday 20 January 2025  06:21:50 -0500 (0:00:00.050)       0:10:19.470 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Monday 20 January 2025  06:21:50 -0500 (0:00:00.064)       0:10:19.535 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Monday 20 January 2025  06:21:50 -0500 (0:00:00.057)       0:10:19.592 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Monday 20 January 2025  06:21:50 -0500 (0:00:00.046)       0:10:19.639 ******** 
ok: [managed-node1] => {
    "changed": false,
    "stat": {
        "atime": 1737372103.775679,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1737372103.775679,
        "dev": 6,
        "device_type": 2048,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 35127,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/blockdevice",
        "mode": "0660",
        "mtime": 1737372103.775679,
        "nlink": 1,
        "path": "/dev/sda",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Monday 20 January 2025  06:21:51 -0500 (0:00:00.474)       0:10:20.113 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Monday 20 January 2025  06:21:51 -0500 (0:00:00.044)       0:10:20.157 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Monday 20 January 2025  06:21:51 -0500 (0:00:00.031)       0:10:20.188 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Monday 20 January 2025  06:21:51 -0500 (0:00:00.026)       0:10:20.215 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "st_volume_type": "disk"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Monday 20 January 2025  06:21:51 -0500 (0:00:00.034)       0:10:20.250 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Monday 20 January 2025  06:21:51 -0500 (0:00:00.033)       0:10:20.283 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Monday 20 January 2025  06:21:51 -0500 (0:00:00.032)       0:10:20.316 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Monday 20 January 2025  06:21:51 -0500 (0:00:00.030)       0:10:20.346 ******** 
ok: [managed-node1] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Monday 20 January 2025  06:21:54 -0500 (0:00:02.963)       0:10:23.309 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Monday 20 January 2025  06:21:54 -0500 (0:00:00.021)       0:10:23.331 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Monday 20 January 2025  06:21:54 -0500 (0:00:00.020)       0:10:23.351 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Monday 20 January 2025  06:21:54 -0500 (0:00:00.015)       0:10:23.367 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Monday 20 January 2025  06:21:54 -0500 (0:00:00.021)       0:10:23.389 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Monday 20 January 2025  06:21:54 -0500 (0:00:00.019)       0:10:23.409 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Monday 20 January 2025  06:21:54 -0500 (0:00:00.015)       0:10:23.424 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Monday 20 January 2025  06:21:54 -0500 (0:00:00.016)       0:10:23.441 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Monday 20 January 2025  06:21:54 -0500 (0:00:00.019)       0:10:23.461 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Monday 20 January 2025  06:21:54 -0500 (0:00:00.034)       0:10:23.495 ******** 
ok: [managed-node1] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Monday 20 January 2025  06:21:54 -0500 (0:00:00.040)       0:10:23.536 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Monday 20 January 2025  06:21:54 -0500 (0:00:00.030)       0:10:23.567 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Monday 20 January 2025  06:21:54 -0500 (0:00:00.036)       0:10:23.604 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Monday 20 January 2025  06:21:54 -0500 (0:00:00.034)       0:10:23.638 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Monday 20 January 2025  06:21:54 -0500 (0:00:00.034)       0:10:23.672 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Monday 20 January 2025  06:21:54 -0500 (0:00:00.035)       0:10:23.708 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Monday 20 January 2025  06:21:54 -0500 (0:00:00.032)       0:10:23.740 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Monday 20 January 2025  06:21:55 -0500 (0:00:00.032)       0:10:23.772 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Monday 20 January 2025  06:21:55 -0500 (0:00:00.039)       0:10:23.812 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Monday 20 January 2025  06:21:55 -0500 (0:00:00.030)       0:10:23.843 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Monday 20 January 2025  06:21:55 -0500 (0:00:00.032)       0:10:23.876 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Monday 20 January 2025  06:21:55 -0500 (0:00:00.032)       0:10:23.908 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Monday 20 January 2025  06:21:55 -0500 (0:00:00.036)       0:10:23.944 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Monday 20 January 2025  06:21:55 -0500 (0:00:00.039)       0:10:23.984 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Monday 20 January 2025  06:21:55 -0500 (0:00:00.029)       0:10:24.014 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Monday 20 January 2025  06:21:55 -0500 (0:00:00.031)       0:10:24.045 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Monday 20 January 2025  06:21:55 -0500 (0:00:00.031)       0:10:24.077 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Monday 20 January 2025  06:21:55 -0500 (0:00:00.027)       0:10:24.104 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Monday 20 January 2025  06:21:55 -0500 (0:00:00.028)       0:10:24.132 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Monday 20 January 2025  06:21:55 -0500 (0:00:00.021)       0:10:24.154 ******** 
skipping: [managed-node1] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Monday 20 January 2025  06:21:55 -0500 (0:00:00.026)       0:10:24.180 ******** 
skipping: [managed-node1] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Monday 20 January 2025  06:21:55 -0500 (0:00:00.019)       0:10:24.200 ******** 
skipping: [managed-node1] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Monday 20 January 2025  06:21:55 -0500 (0:00:00.027)       0:10:24.228 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Monday 20 January 2025  06:21:55 -0500 (0:00:00.021)       0:10:24.249 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Monday 20 January 2025  06:21:55 -0500 (0:00:00.022)       0:10:24.272 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Monday 20 January 2025  06:21:55 -0500 (0:00:00.022)       0:10:24.294 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Monday 20 January 2025  06:21:55 -0500 (0:00:00.019)       0:10:24.314 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Monday 20 January 2025  06:21:55 -0500 (0:00:00.020)       0:10:24.334 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Monday 20 January 2025  06:21:55 -0500 (0:00:00.019)       0:10:24.354 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Monday 20 January 2025  06:21:55 -0500 (0:00:00.019)       0:10:24.374 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Monday 20 January 2025  06:21:55 -0500 (0:00:00.021)       0:10:24.395 ******** 
skipping: [managed-node1] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Monday 20 January 2025  06:21:55 -0500 (0:00:00.022)       0:10:24.418 ******** 
skipping: [managed-node1] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Monday 20 January 2025  06:21:55 -0500 (0:00:00.019)       0:10:24.438 ******** 
skipping: [managed-node1] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Monday 20 January 2025  06:21:55 -0500 (0:00:00.019)       0:10:24.457 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Monday 20 January 2025  06:21:55 -0500 (0:00:00.020)       0:10:24.478 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Monday 20 January 2025  06:21:55 -0500 (0:00:00.019)       0:10:24.497 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Monday 20 January 2025  06:21:55 -0500 (0:00:00.021)       0:10:24.519 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Monday 20 January 2025  06:21:55 -0500 (0:00:00.019)       0:10:24.538 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Monday 20 January 2025  06:21:55 -0500 (0:00:00.020)       0:10:24.558 ******** 
ok: [managed-node1] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Monday 20 January 2025  06:21:55 -0500 (0:00:00.023)       0:10:24.582 ******** 
ok: [managed-node1] => {
    "storage_test_expected_size": "4294967296"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Monday 20 January 2025  06:21:55 -0500 (0:00:00.021)       0:10:24.603 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Monday 20 January 2025  06:21:55 -0500 (0:00:00.020)       0:10:24.624 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Monday 20 January 2025  06:21:55 -0500 (0:00:00.022)       0:10:24.647 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Monday 20 January 2025  06:21:55 -0500 (0:00:00.021)       0:10:24.668 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Monday 20 January 2025  06:21:55 -0500 (0:00:00.020)       0:10:24.689 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Monday 20 January 2025  06:21:55 -0500 (0:00:00.021)       0:10:24.711 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Monday 20 January 2025  06:21:55 -0500 (0:00:00.022)       0:10:24.733 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Monday 20 January 2025  06:21:55 -0500 (0:00:00.021)       0:10:24.755 ******** 
skipping: [managed-node1] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Monday 20 January 2025  06:21:56 -0500 (0:00:00.065)       0:10:24.821 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Monday 20 January 2025  06:21:56 -0500 (0:00:00.021)       0:10:24.843 ******** 
ok: [managed-node1] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}
META: ran handlers
META: ran handlers

PLAY RECAP *********************************************************************
managed-node1              : ok=1229 changed=60   unreachable=0    failed=9    skipped=1061 rescued=9    ignored=0   

Monday 20 January 2025  06:21:56 -0500 (0:00:00.014)       0:10:24.857 ******** 
=============================================================================== 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state -- 15.26s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state -- 14.82s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state -- 14.71s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state -- 14.65s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state -- 14.56s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state -- 14.30s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 5.05s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.70s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.57s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Get required packages --------------- 4.53s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.45s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Get required packages --------------- 4.44s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 
fedora.linux_system_roles.storage : Get required packages --------------- 4.36s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.33s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Get required packages --------------- 4.32s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 
fedora.linux_system_roles.storage : Get required packages --------------- 4.30s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 
fedora.linux_system_roles.storage : Get required packages --------------- 4.29s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.28s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.24s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Get required packages --------------- 4.23s
/tmp/collections-fdQ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19