ansible-playbook [core 2.12.6]
  config file = /etc/ansible/ansible.cfg
  configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
  ansible python module location = /usr/lib/python3.9/site-packages/ansible
  ansible collection location = /tmp/tmp5bkr4li_
  executable location = /usr/bin/ansible-playbook
  python version = 3.9.13 (main, May 18 2022, 00:00:00) [GCC 11.3.1 20220421 (Red Hat 11.3.1-2)]
  jinja version = 2.11.3
  libyaml = True
Using /etc/ansible/ansible.cfg as config file
Skipping callback 'debug', as we already have a stdout callback.
Skipping callback 'default', as we already have a stdout callback.
Skipping callback 'minimal', as we already have a stdout callback.
Skipping callback 'oneline', as we already have a stdout callback.

PLAYBOOK: rhel-7_setup.yml *****************************************************
1 plays in /cache/rhel-7_setup.yml

PLAY [Setup repos] *************************************************************
META: ran handlers

TASK [set up internal repositories] ********************************************
task path: /cache/rhel-7_setup.yml:5
Thursday 21 July 2022  15:43:01 +0000 (0:00:00.018)       0:00:00.018 ********* 
changed: [/cache/rhel-7.qcow2] => (item=None) => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
changed: [/cache/rhel-7.qcow2] => (item=None) => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
changed: [/cache/rhel-7.qcow2] => (item=None) => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
changed: [/cache/rhel-7.qcow2] => (item=None) => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
changed: [/cache/rhel-7.qcow2] => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
META: ran handlers
META: ran handlers

PLAY RECAP *********************************************************************
/cache/rhel-7.qcow2        : ok=1    changed=1    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   

Thursday 21 July 2022  15:43:03 +0000 (0:00:01.435)       0:00:01.454 ********* 
=============================================================================== 
set up internal repositories -------------------------------------------- 1.44s
/cache/rhel-7_setup.yml:5 -----------------------------------------------------

PLAYBOOK: tests_misc_scsi_generated.yml ****************************************
2 plays in /tmp/tmpaxjje44y/tests/tests_misc_scsi_generated.yml

PLAY [all] *********************************************************************

TASK [Gathering Facts] *********************************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc_scsi_generated.yml:3
Thursday 21 July 2022  15:43:03 +0000 (0:00:00.021)       0:00:01.475 ********* 
ok: [/cache/rhel-7.qcow2]
META: ran handlers

TASK [set disk interface for test] *********************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc_scsi_generated.yml:7
Thursday 21 July 2022  15:43:04 +0000 (0:00:01.038)       0:00:02.514 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_use_interface": "scsi"
    },
    "changed": false
}
META: ran handlers
META: ran handlers

PLAY [all] *********************************************************************

TASK [Gathering Facts] *********************************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:2
Thursday 21 July 2022  15:43:04 +0000 (0:00:00.049)       0:00:02.564 ********* 
ok: [/cache/rhel-7.qcow2]
META: ran handlers

TASK [include_role : linux-system-roles.storage] *******************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:16
Thursday 21 July 2022  15:43:05 +0000 (0:00:00.749)       0:00:03.313 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:43:05 +0000 (0:00:00.033)       0:00:03.347 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:43:05 +0000 (0:00:00.029)       0:00:03.376 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:43:05 +0000 (0:00:00.411)       0:00:03.788 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:43:05 +0000 (0:00:00.070)       0:00:03.859 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:43:05 +0000 (0:00:00.030)       0:00:03.889 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:43:05 +0000 (0:00:00.028)       0:00:03.918 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:43:05 +0000 (0:00:00.044)       0:00:03.962 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:43:05 +0000 (0:00:00.016)       0:00:03.979 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "changed": true,
    "changes": {
        "installed": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "rc": 0,
    "results": [
        "Loaded plugins: search-disabled-repos\nResolving Dependencies\n--> Running transaction check\n---> Package libblockdev-crypto.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libblockdev-utils(x86-64) = 2.18-5.el7 for package: libblockdev-crypto-2.18-5.el7.x86_64\n--> Processing Dependency: libvolume_key.so.1()(64bit) for package: libblockdev-crypto-2.18-5.el7.x86_64\n--> Processing Dependency: libbd_utils.so.2()(64bit) for package: libblockdev-crypto-2.18-5.el7.x86_64\n---> Package libblockdev-dm.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libdmraid.so.1(Base)(64bit) for package: libblockdev-dm-2.18-5.el7.x86_64\n--> Processing Dependency: dmraid for package: libblockdev-dm-2.18-5.el7.x86_64\n--> Processing Dependency: libdmraid.so.1()(64bit) for package: libblockdev-dm-2.18-5.el7.x86_64\n---> Package libblockdev-lvm.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: lvm2 for package: libblockdev-lvm-2.18-5.el7.x86_64\n--> Processing Dependency: device-mapper-persistent-data for package: libblockdev-lvm-2.18-5.el7.x86_64\n---> Package libblockdev-mdraid.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: mdadm for package: libblockdev-mdraid-2.18-5.el7.x86_64\n--> Processing Dependency: libbytesize.so.1()(64bit) for package: libblockdev-mdraid-2.18-5.el7.x86_64\n---> Package libblockdev-swap.x86_64 0:2.18-5.el7 will be installed\n---> Package python-enum34.noarch 0:1.0.4-1.el7 will be installed\n---> Package python2-blivet3.noarch 1:3.1.3-3.el7 will be installed\n--> Processing Dependency: blivet3-data = 1:3.1.3-3.el7 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-bytesize >= 0.3 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-blockdev >= 2.17 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: pyparted >= 3.9 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-hawkey for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: lsof for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Running transaction check\n---> Package blivet3-data.noarch 1:3.1.3-3.el7 will be installed\n---> Package device-mapper-persistent-data.x86_64 0:0.8.5-3.el7_9.2 will be installed\n--> Processing Dependency: libaio.so.1(LIBAIO_0.4)(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n--> Processing Dependency: libaio.so.1(LIBAIO_0.1)(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n--> Processing Dependency: libaio.so.1()(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n---> Package dmraid.x86_64 0:1.0.0.rc16-28.el7 will be installed\n--> Processing Dependency: libdevmapper-event.so.1.02(Base)(64bit) for package: dmraid-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: dmraid-events for package: dmraid-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: libdevmapper-event.so.1.02()(64bit) for package: dmraid-1.0.0.rc16-28.el7.x86_64\n---> Package libblockdev-utils.x86_64 0:2.18-5.el7 will be installed\n---> Package libbytesize.x86_64 0:1.2-1.el7 will be installed\n--> Processing Dependency: libmpfr.so.4()(64bit) for package: libbytesize-1.2-1.el7.x86_64\n---> Package lsof.x86_64 0:4.87-6.el7 will be installed\n---> Package lvm2.x86_64 7:2.02.187-6.el7_9.5 will be installed\n--> Processing Dependency: lvm2-libs = 7:2.02.187-6.el7_9.5 for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n--> Processing Dependency: liblvm2app.so.2.2(Base)(64bit) for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n--> Processing Dependency: liblvm2app.so.2.2()(64bit) for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n---> Package mdadm.x86_64 0:4.1-9.el7_9 will be installed\n--> Processing Dependency: libreport-filesystem for package: mdadm-4.1-9.el7_9.x86_64\n---> Package pyparted.x86_64 1:3.9-15.el7 will be installed\n---> Package python2-blockdev.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libblockdev(x86-64) = 2.18-5.el7 for package: python2-blockdev-2.18-5.el7.x86_64\n---> Package python2-bytesize.x86_64 0:1.2-1.el7 will be installed\n---> Package python2-hawkey.x86_64 0:0.22.5-2.el7_9 will be installed\n--> Processing Dependency: libdnf(x86-64) = 0.22.5-2.el7_9 for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: python2-libdnf = 0.22.5-2.el7_9 for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolv.so.0(SOLV_1.0)(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolvext.so.0(SOLV_1.0)(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libdnf.so.2()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libjson-glib-1.0.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libmodulemd.so.1()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: librepo.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: librhsm.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolv.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolvext.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n---> Package volume_key-libs.x86_64 0:0.3.9-9.el7 will be installed\n--> Running transaction check\n---> Package device-mapper-event-libs.x86_64 7:1.02.170-6.el7_9.5 will be installed\n---> Package dmraid-events.x86_64 0:1.0.0.rc16-28.el7 will be installed\n--> Processing Dependency: sgpio for package: dmraid-events-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: device-mapper-event for package: dmraid-events-1.0.0.rc16-28.el7.x86_64\n---> Package json-glib.x86_64 0:1.4.2-2.el7 will be installed\n---> Package libaio.x86_64 0:0.3.109-13.el7 will be installed\n---> Package libblockdev.x86_64 0:2.18-5.el7 will be installed\n---> Package libdnf.x86_64 0:0.22.5-2.el7_9 will be installed\n---> Package libmodulemd.x86_64 0:1.6.3-1.el7 will be installed\n---> Package librepo.x86_64 0:1.8.1-8.el7_9 will be installed\n---> Package libreport-filesystem.x86_64 0:2.1.11-53.el7 will be installed\n---> Package librhsm.x86_64 0:0.0.3-3.el7_9 will be installed\n---> Package libsolv.x86_64 0:0.6.34-4.el7 will be installed\n---> Package lvm2-libs.x86_64 7:2.02.187-6.el7_9.5 will be installed\n---> Package mpfr.x86_64 0:3.1.1-4.el7 will be installed\n---> Package python2-libdnf.x86_64 0:0.22.5-2.el7_9 will be installed\n--> Running transaction check\n---> Package device-mapper-event.x86_64 7:1.02.170-6.el7_9.5 will be installed\n---> Package sgpio.x86_64 0:1.2.0.10-13.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package            Arch   Version              Repository                 Size\n================================================================================\nInstalling:\n libblockdev-crypto x86_64 2.18-5.el7           rhel                       60 k\n libblockdev-dm     x86_64 2.18-5.el7           rhel-optional              54 k\n libblockdev-lvm    x86_64 2.18-5.el7           rhel                       62 k\n libblockdev-mdraid x86_64 2.18-5.el7           rhel                       57 k\n libblockdev-swap   x86_64 2.18-5.el7           rhel                       52 k\n python-enum34      noarch 1.0.4-1.el7          rhel                       52 k\n python2-blivet3    noarch 1:3.1.3-3.el7        rhel                      851 k\nInstalling for dependencies:\n blivet3-data       noarch 1:3.1.3-3.el7        rhel                       77 k\n device-mapper-event\n                    x86_64 7:1.02.170-6.el7_9.5 rhel                      192 k\n device-mapper-event-libs\n                    x86_64 7:1.02.170-6.el7_9.5 rhel                      192 k\n device-mapper-persistent-data\n                    x86_64 0.8.5-3.el7_9.2      rhel                      423 k\n dmraid             x86_64 1.0.0.rc16-28.el7    rhel                      151 k\n dmraid-events      x86_64 1.0.0.rc16-28.el7    rhel                       21 k\n json-glib          x86_64 1.4.2-2.el7          rhel                      134 k\n libaio             x86_64 0.3.109-13.el7       rhel                       24 k\n libblockdev        x86_64 2.18-5.el7           rhel                      119 k\n libblockdev-utils  x86_64 2.18-5.el7           rhel                       59 k\n libbytesize        x86_64 1.2-1.el7            rhel                       52 k\n libdnf             x86_64 0.22.5-2.el7_9       rhel-7-server-extras-rpms 536 k\n libmodulemd        x86_64 1.6.3-1.el7          rhel-7-server-extras-rpms 153 k\n librepo            x86_64 1.8.1-8.el7_9        rhel                       82 k\n libreport-filesystem\n                    x86_64 2.1.11-53.el7        rhel                       41 k\n librhsm            x86_64 0.0.3-3.el7_9        rhel-7-server-extras-rpms  28 k\n libsolv            x86_64 0.6.34-4.el7         rhel                      329 k\n lsof               x86_64 4.87-6.el7           rhel                      331 k\n lvm2               x86_64 7:2.02.187-6.el7_9.5 rhel                      1.3 M\n lvm2-libs          x86_64 7:2.02.187-6.el7_9.5 rhel                      1.1 M\n mdadm              x86_64 4.1-9.el7_9          rhel                      440 k\n mpfr               x86_64 3.1.1-4.el7          rhel                      203 k\n pyparted           x86_64 1:3.9-15.el7         rhel                      195 k\n python2-blockdev   x86_64 2.18-5.el7           rhel                       61 k\n python2-bytesize   x86_64 1.2-1.el7            rhel                       22 k\n python2-hawkey     x86_64 0.22.5-2.el7_9       rhel-7-server-extras-rpms  71 k\n python2-libdnf     x86_64 0.22.5-2.el7_9       rhel-7-server-extras-rpms 611 k\n sgpio              x86_64 1.2.0.10-13.el7      rhel                       14 k\n volume_key-libs    x86_64 0.3.9-9.el7          rhel                      141 k\n\nTransaction Summary\n================================================================================\nInstall  7 Packages (+29 Dependent packages)\n\nTotal download size: 8.2 M\nInstalled size: 24 M\nDownloading packages:\n--------------------------------------------------------------------------------\nTotal                                               20 MB/s | 8.2 MB  00:00     \nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n  Installing : libblockdev-utils-2.18-5.el7.x86_64                         1/36 \n  Installing : 7:device-mapper-event-libs-1.02.170-6.el7_9.5.x86_64        2/36 \n  Installing : json-glib-1.4.2-2.el7.x86_64                                3/36 \n  Installing : librhsm-0.0.3-3.el7_9.x86_64                                4/36 \n  Installing : libsolv-0.6.34-4.el7.x86_64                                 5/36 \n  Installing : libaio-0.3.109-13.el7.x86_64                                6/36 \n  Installing : librepo-1.8.1-8.el7_9.x86_64                                7/36 \n  Installing : libmodulemd-1.6.3-1.el7.x86_64                              8/36 \n  Installing : libdnf-0.22.5-2.el7_9.x86_64                                9/36 \n  Installing : device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64       10/36 \n  Installing : 7:device-mapper-event-1.02.170-6.el7_9.5.x86_64            11/36 \n  Installing : 7:lvm2-libs-2.02.187-6.el7_9.5.x86_64                      12/36 \n  Installing : 7:lvm2-2.02.187-6.el7_9.5.x86_64                           13/36 \n  Installing : python2-libdnf-0.22.5-2.el7_9.x86_64                       14/36 \n  Installing : python2-hawkey-0.22.5-2.el7_9.x86_64                       15/36 \n  Installing : libblockdev-2.18-5.el7.x86_64                              16/36 \n  Installing : python2-blockdev-2.18-5.el7.x86_64                         17/36 \n  Installing : 1:pyparted-3.9-15.el7.x86_64                               18/36 \n  Installing : sgpio-1.2.0.10-13.el7.x86_64                               19/36 \n  Installing : dmraid-1.0.0.rc16-28.el7.x86_64                            20/36 \n  Installing : dmraid-events-1.0.0.rc16-28.el7.x86_64                     21/36 \n  Installing : volume_key-libs-0.3.9-9.el7.x86_64                         22/36 \n  Installing : mpfr-3.1.1-4.el7.x86_64                                    23/36 \n  Installing : libbytesize-1.2-1.el7.x86_64                               24/36 \n  Installing : python2-bytesize-1.2-1.el7.x86_64                          25/36 \n  Installing : libreport-filesystem-2.1.11-53.el7.x86_64                  26/36 \n  Installing : mdadm-4.1-9.el7_9.x86_64                                   27/36 \n  Installing : 1:blivet3-data-3.1.3-3.el7.noarch                          28/36 \n  Installing : lsof-4.87-6.el7.x86_64                                     29/36 \n  Installing : 1:python2-blivet3-3.1.3-3.el7.noarch                       30/36 \n  Installing : libblockdev-mdraid-2.18-5.el7.x86_64                       31/36 \n  Installing : libblockdev-crypto-2.18-5.el7.x86_64                       32/36 \n  Installing : libblockdev-dm-2.18-5.el7.x86_64                           33/36 \n  Installing : libblockdev-lvm-2.18-5.el7.x86_64                          34/36 \n  Installing : libblockdev-swap-2.18-5.el7.x86_64                         35/36 \n  Installing : python-enum34-1.0.4-1.el7.noarch                           36/36 \n  Verifying  : 7:device-mapper-event-1.02.170-6.el7_9.5.x86_64             1/36 \n  Verifying  : libblockdev-swap-2.18-5.el7.x86_64                          2/36 \n  Verifying  : librhsm-0.0.3-3.el7_9.x86_64                                3/36 \n  Verifying  : libblockdev-lvm-2.18-5.el7.x86_64                           4/36 \n  Verifying  : lsof-4.87-6.el7.x86_64                                      5/36 \n  Verifying  : libblockdev-mdraid-2.18-5.el7.x86_64                        6/36 \n  Verifying  : libdnf-0.22.5-2.el7_9.x86_64                                7/36 \n  Verifying  : python-enum34-1.0.4-1.el7.noarch                            8/36 \n  Verifying  : 1:blivet3-data-3.1.3-3.el7.noarch                           9/36 \n  Verifying  : dmraid-events-1.0.0.rc16-28.el7.x86_64                     10/36 \n  Verifying  : python2-blockdev-2.18-5.el7.x86_64                         11/36 \n  Verifying  : libmodulemd-1.6.3-1.el7.x86_64                             12/36 \n  Verifying  : librepo-1.8.1-8.el7_9.x86_64                               13/36 \n  Verifying  : libblockdev-dm-2.18-5.el7.x86_64                           14/36 \n  Verifying  : json-glib-1.4.2-2.el7.x86_64                               15/36 \n  Verifying  : libaio-0.3.109-13.el7.x86_64                               16/36 \n  Verifying  : 7:lvm2-libs-2.02.187-6.el7_9.5.x86_64                      17/36 \n  Verifying  : python2-hawkey-0.22.5-2.el7_9.x86_64                       18/36 \n  Verifying  : python2-bytesize-1.2-1.el7.x86_64                          19/36 \n  Verifying  : libblockdev-2.18-5.el7.x86_64                              20/36 \n  Verifying  : libreport-filesystem-2.1.11-53.el7.x86_64                  21/36 \n  Verifying  : libbytesize-1.2-1.el7.x86_64                               22/36 \n  Verifying  : 7:device-mapper-event-libs-1.02.170-6.el7_9.5.x86_64       23/36 \n  Verifying  : python2-libdnf-0.22.5-2.el7_9.x86_64                       24/36 \n  Verifying  : 7:lvm2-2.02.187-6.el7_9.5.x86_64                           25/36 \n  Verifying  : libblockdev-utils-2.18-5.el7.x86_64                        26/36 \n  Verifying  : mpfr-3.1.1-4.el7.x86_64                                    27/36 \n  Verifying  : volume_key-libs-0.3.9-9.el7.x86_64                         28/36 \n  Verifying  : libsolv-0.6.34-4.el7.x86_64                                29/36 \n  Verifying  : device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64       30/36 \n  Verifying  : 1:python2-blivet3-3.1.3-3.el7.noarch                       31/36 \n  Verifying  : dmraid-1.0.0.rc16-28.el7.x86_64                            32/36 \n  Verifying  : mdadm-4.1-9.el7_9.x86_64                                   33/36 \n  Verifying  : sgpio-1.2.0.10-13.el7.x86_64                               34/36 \n  Verifying  : libblockdev-crypto-2.18-5.el7.x86_64                       35/36 \n  Verifying  : 1:pyparted-3.9-15.el7.x86_64                               36/36 \n\nInstalled:\n  libblockdev-crypto.x86_64 0:2.18-5.el7 libblockdev-dm.x86_64 0:2.18-5.el7    \n  libblockdev-lvm.x86_64 0:2.18-5.el7    libblockdev-mdraid.x86_64 0:2.18-5.el7\n  libblockdev-swap.x86_64 0:2.18-5.el7   python-enum34.noarch 0:1.0.4-1.el7    \n  python2-blivet3.noarch 1:3.1.3-3.el7  \n\nDependency Installed:\n  blivet3-data.noarch 1:3.1.3-3.el7                                             \n  device-mapper-event.x86_64 7:1.02.170-6.el7_9.5                               \n  device-mapper-event-libs.x86_64 7:1.02.170-6.el7_9.5                          \n  device-mapper-persistent-data.x86_64 0:0.8.5-3.el7_9.2                        \n  dmraid.x86_64 0:1.0.0.rc16-28.el7                                             \n  dmraid-events.x86_64 0:1.0.0.rc16-28.el7                                      \n  json-glib.x86_64 0:1.4.2-2.el7                                                \n  libaio.x86_64 0:0.3.109-13.el7                                                \n  libblockdev.x86_64 0:2.18-5.el7                                               \n  libblockdev-utils.x86_64 0:2.18-5.el7                                         \n  libbytesize.x86_64 0:1.2-1.el7                                                \n  libdnf.x86_64 0:0.22.5-2.el7_9                                                \n  libmodulemd.x86_64 0:1.6.3-1.el7                                              \n  librepo.x86_64 0:1.8.1-8.el7_9                                                \n  libreport-filesystem.x86_64 0:2.1.11-53.el7                                   \n  librhsm.x86_64 0:0.0.3-3.el7_9                                                \n  libsolv.x86_64 0:0.6.34-4.el7                                                 \n  lsof.x86_64 0:4.87-6.el7                                                      \n  lvm2.x86_64 7:2.02.187-6.el7_9.5                                              \n  lvm2-libs.x86_64 7:2.02.187-6.el7_9.5                                         \n  mdadm.x86_64 0:4.1-9.el7_9                                                    \n  mpfr.x86_64 0:3.1.1-4.el7                                                     \n  pyparted.x86_64 1:3.9-15.el7                                                  \n  python2-blockdev.x86_64 0:2.18-5.el7                                          \n  python2-bytesize.x86_64 0:1.2-1.el7                                           \n  python2-hawkey.x86_64 0:0.22.5-2.el7_9                                        \n  python2-libdnf.x86_64 0:0.22.5-2.el7_9                                        \n  sgpio.x86_64 0:1.2.0.10-13.el7                                                \n  volume_key-libs.x86_64 0:0.3.9-9.el7                                          \n\nComplete!\n"
    ]
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:43:14 +0000 (0:00:09.123)       0:00:13.102 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:43:14 +0000 (0:00:00.063)       0:00:13.166 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:43:14 +0000 (0:00:00.069)       0:00:13.236 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:43:15 +0000 (0:00:00.598)       0:00:13.834 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : check if the COPR support packages should be installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2
Thursday 21 July 2022  15:43:15 +0000 (0:00:00.042)       0:00:13.876 ********* 

TASK [linux-system-roles.storage : make sure COPR support packages are present] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13
Thursday 21 July 2022  15:43:15 +0000 (0:00:00.031)       0:00:13.908 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable COPRs] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18
Thursday 21 July 2022  15:43:15 +0000 (0:00:00.032)       0:00:13.940 ********* 

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:43:15 +0000 (0:00:00.031)       0:00:13.971 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "rc": 0,
    "results": []
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:43:16 +0000 (0:00:00.546)       0:00:14.518 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "arp-ethers.service": {
                "name": "arp-ethers.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "brandbot.service": {
                "name": "brandbot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "console-shell.service": {
                "name": "console-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "display-manager.service": {
                "name": "display-manager.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dmraid-activation.service": {
                "name": "dmraid-activation.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "exim.service": {
                "name": "exim.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "active"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ip6tables.service": {
                "name": "ip6tables.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "iptables.service": {
                "name": "iptables.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-activation-early.service": {
                "name": "lvm2-activation-early.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "lvm2-activation.service": {
                "name": "lvm2-activation.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "netconsole": {
                "name": "netconsole",
                "source": "sysv",
                "state": "stopped",
                "status": "disabled"
            },
            "network": {
                "name": "network",
                "source": "sysv",
                "state": "running",
                "status": "enabled"
            },
            "network.service": {
                "name": "network.service",
                "source": "systemd",
                "state": "stopped",
                "status": "active"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-config.service": {
                "name": "nfs-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-idmap.service": {
                "name": "nfs-idmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-lock.service": {
                "name": "nfs-lock.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-secure-server.service": {
                "name": "nfs-secure-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "nfs-secure.service": {
                "name": "nfs-secure.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs.service": {
                "name": "nfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfslock.service": {
                "name": "nfslock.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "ntpd.service": {
                "name": "ntpd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "ntpdate.service": {
                "name": "ntpdate.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "ovirt-guest-agent.service": {
                "name": "ovirt-guest-agent.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "postfix.service": {
                "name": "postfix.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rhcd.service": {
                "name": "rhcd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-configure.service": {
                "name": "rhel-configure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-domainname.service": {
                "name": "rhel-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-import-state.service": {
                "name": "rhel-import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-readonly.service": {
                "name": "rhel-readonly.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhnsd": {
                "name": "rhnsd",
                "source": "sysv",
                "state": "running",
                "status": "enabled"
            },
            "rhnsd.service": {
                "name": "rhnsd.service",
                "source": "systemd",
                "state": "running",
                "status": "active"
            },
            "rhsm-facts.service": {
                "name": "rhsm-facts.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rhsm.service": {
                "name": "rhsm.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rhsmcertd.service": {
                "name": "rhsmcertd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpcgssd.service": {
                "name": "rpcgssd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rpcidmapd.service": {
                "name": "rpcidmapd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rsyncd.service": {
                "name": "rsyncd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rsyncd@.service": {
                "name": "rsyncd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service",
                "source": "systemd",
                "state": "stopped",
                "status": "active"
            },
            "sendmail.service": {
                "name": "sendmail.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service",
                "source": "systemd",
                "state": "running",
                "status": "active"
            },
            "sntp.service": {
                "name": "sntp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "sshd-keygen.service": {
                "name": "sshd-keygen.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-importd.service": {
                "name": "systemd-importd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-machined.service": {
                "name": "systemd-machined.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-timesyncd.service": {
                "name": "systemd-timesyncd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ypbind.service": {
                "name": "ypbind.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "yppasswdd.service": {
                "name": "yppasswdd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "ypserv.service": {
                "name": "ypserv.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "ypxfrd.service": {
                "name": "ypxfrd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            }
        }
    },
    "changed": false
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:43:17 +0000 (0:00:01.185)       0:00:15.703 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:43:17 +0000 (0:00:00.055)       0:00:15.758 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:43:17 +0000 (0:00:00.020)       0:00:15.779 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91
Thursday 21 July 2022  15:43:18 +0000 (0:00:00.475)       0:00:16.255 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:43:18 +0000 (0:00:00.036)       0:00:16.292 ********* 

TASK [linux-system-roles.storage : show blivet_output] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109
Thursday 21 July 2022  15:43:18 +0000 (0:00:00.019)       0:00:16.311 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [],
        "changed": false,
        "crypts": [],
        "failed": false,
        "leaves": [],
        "mounts": [],
        "packages": [],
        "pools": [],
        "volumes": []
    }
}

TASK [linux-system-roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114
Thursday 21 July 2022  15:43:18 +0000 (0:00:00.035)       0:00:16.346 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118
Thursday 21 July 2022  15:43:18 +0000 (0:00:00.039)       0:00:16.386 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : remove obsolete mounts] *********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134
Thursday 21 July 2022  15:43:18 +0000 (0:00:00.035)       0:00:16.421 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146
Thursday 21 July 2022  15:43:18 +0000 (0:00:00.031)       0:00:16.453 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : set up new/current mounts] ******************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151
Thursday 21 July 2022  15:43:18 +0000 (0:00:00.024)       0:00:16.477 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163
Thursday 21 July 2022  15:43:18 +0000 (0:00:00.033)       0:00:16.511 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171
Thursday 21 July 2022  15:43:18 +0000 (0:00:00.023)       0:00:16.535 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658418192.4109497,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176
Thursday 21 July 2022  15:43:18 +0000 (0:00:00.443)       0:00:16.978 ********* 

TASK [linux-system-roles.storage : Update facts] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198
Thursday 21 July 2022  15:43:18 +0000 (0:00:00.022)       0:00:17.001 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [Mark tasks to be skipped] ************************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:19
Thursday 21 July 2022  15:43:19 +0000 (0:00:00.834)       0:00:17.835 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_skip_checks": [
            "blivet_available",
            "packages_installed",
            "service_facts"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:26
Thursday 21 July 2022  15:43:19 +0000 (0:00:00.062)       0:00:17.898 ********* 
included: /tmp/tmpaxjje44y/tests/get_unused_disk.yml for /cache/rhel-7.qcow2

TASK [Find unused disks in the system] *****************************************
task path: /tmp/tmpaxjje44y/tests/get_unused_disk.yml:2
Thursday 21 July 2022  15:43:19 +0000 (0:00:00.034)       0:00:17.933 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "disks": [
        "sdb"
    ]
}

TASK [Set unused_disks if necessary] *******************************************
task path: /tmp/tmpaxjje44y/tests/get_unused_disk.yml:9
Thursday 21 July 2022  15:43:20 +0000 (0:00:00.491)       0:00:18.424 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "unused_disks": [
            "sdb"
        ]
    },
    "changed": false
}

TASK [Exit playbook when there's not enough unused disks in the system] ********
task path: /tmp/tmpaxjje44y/tests/get_unused_disk.yml:14
Thursday 21 July 2022  15:43:20 +0000 (0:00:00.037)       0:00:18.461 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Print unused disks] ******************************************************
task path: /tmp/tmpaxjje44y/tests/get_unused_disk.yml:19
Thursday 21 July 2022  15:43:20 +0000 (0:00:00.038)       0:00:18.500 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "unused_disks": [
        "sdb"
    ]
}

TASK [Test creating ext4 filesystem with valid parameter "-Fb 4096"] ***********
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:31
Thursday 21 July 2022  15:43:20 +0000 (0:00:00.062)       0:00:18.563 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:43:20 +0000 (0:00:00.069)       0:00:18.632 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:43:20 +0000 (0:00:00.035)       0:00:18.668 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:43:20 +0000 (0:00:00.412)       0:00:19.081 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:43:20 +0000 (0:00:00.058)       0:00:19.139 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:43:20 +0000 (0:00:00.030)       0:00:19.170 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:43:20 +0000 (0:00:00.031)       0:00:19.201 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:43:21 +0000 (0:00:00.047)       0:00:19.248 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:43:21 +0000 (0:00:00.018)       0:00:19.267 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:43:21 +0000 (0:00:00.027)       0:00:19.295 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "sdb"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_create_options": "-Fb 4096",
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:43:21 +0000 (0:00:00.038)       0:00:19.333 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:43:21 +0000 (0:00:00.035)       0:00:19.369 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:43:21 +0000 (0:00:00.028)       0:00:19.397 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:43:21 +0000 (0:00:00.027)       0:00:19.425 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:43:21 +0000 (0:00:00.027)       0:00:19.453 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:43:21 +0000 (0:00:00.030)       0:00:19.484 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:43:21 +0000 (0:00:00.056)       0:00:19.540 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:43:21 +0000 (0:00:00.021)       0:00:19.561 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "create format",
            "device": "/dev/sdb",
            "fs_type": "lvmpv"
        },
        {
            "action": "create device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "create device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd",
        "/dev/mapper/foo-test1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext4",
            "opts": "defaults",
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs",
        "lvm2"
    ],
    "pools": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "-Fb 4096",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91
Thursday 21 July 2022  15:43:23 +0000 (0:00:01.687)       0:00:21.249 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:43:23 +0000 (0:00:00.036)       0:00:21.285 ********* 

TASK [linux-system-roles.storage : show blivet_output] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109
Thursday 21 July 2022  15:43:23 +0000 (0:00:00.021)       0:00:21.306 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "create format",
                "device": "/dev/sdb",
                "fs_type": "lvmpv"
            },
            {
                "action": "create device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "create device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd",
            "/dev/mapper/foo-test1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext4",
                "opts": "defaults",
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs",
            "lvm2"
        ],
        "pools": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "-Fb 4096",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [linux-system-roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114
Thursday 21 July 2022  15:43:23 +0000 (0:00:00.039)       0:00:21.346 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "-Fb 4096",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [linux-system-roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118
Thursday 21 July 2022  15:43:23 +0000 (0:00:00.037)       0:00:21.384 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : remove obsolete mounts] *********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134
Thursday 21 July 2022  15:43:23 +0000 (0:00:00.071)       0:00:21.455 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146
Thursday 21 July 2022  15:43:23 +0000 (0:00:00.038)       0:00:21.494 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : set up new/current mounts] ******************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151
Thursday 21 July 2022  15:43:24 +0000 (0:00:00.750)       0:00:22.245 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'ext4', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "opts": "defaults",
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163
Thursday 21 July 2022  15:43:24 +0000 (0:00:00.480)       0:00:22.726 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171
Thursday 21 July 2022  15:43:24 +0000 (0:00:00.456)       0:00:23.182 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658418192.4109497,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176
Thursday 21 July 2022  15:43:25 +0000 (0:00:00.318)       0:00:23.500 ********* 

TASK [linux-system-roles.storage : Update facts] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198
Thursday 21 July 2022  15:43:25 +0000 (0:00:00.023)       0:00:23.524 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:46
Thursday 21 July 2022  15:43:26 +0000 (0:00:00.833)       0:00:24.357 ********* 
included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1
Thursday 21 July 2022  15:43:26 +0000 (0:00:00.039)       0:00:24.397 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "-Fb 4096",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6
Thursday 21 July 2022  15:43:26 +0000 (0:00:00.047)       0:00:24.444 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14
Thursday 21 July 2022  15:43:26 +0000 (0:00:00.032)       0:00:24.477 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/mapper/foo-test1": {
            "fstype": "ext4",
            "label": "",
            "name": "/dev/mapper/foo-test1",
            "size": "4G",
            "type": "lvm",
            "uuid": "9ac0b734-2f82-4721-a26a-c5ea2621165e"
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "LVM2_member",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": "awYqdL-Jqbq-vxDT-PaGG-s0IO-Exqj-zbEdEs"
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-15-42-49-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19
Thursday 21 July 2022  15:43:26 +0000 (0:00:00.439)       0:00:24.916 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003595",
    "end": "2022-07-21 11:43:26.931560",
    "rc": 0,
    "start": "2022-07-21 11:43:26.927965"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0
/dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24
Thursday 21 July 2022  15:43:27 +0000 (0:00:00.442)       0:00:25.359 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003603",
    "end": "2022-07-21 11:43:27.250486",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 11:43:27.246883"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33
Thursday 21 July 2022  15:43:27 +0000 (0:00:00.319)       0:00:25.678 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None})

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:5
Thursday 21 July 2022  15:43:27 +0000 (0:00:00.058)       0:00:25.737 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:18
Thursday 21 July 2022  15:43:27 +0000 (0:00:00.066)       0:00:25.803 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members)
included: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes)

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:1
Thursday 21 July 2022  15:43:27 +0000 (0:00:00.047)       0:00:25.851 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sdb"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:10
Thursday 21 July 2022  15:43:27 +0000 (0:00:00.084)       0:00:25.936 ********* 
ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sdb",
    "pv": "/dev/sdb"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:19
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.441)       0:00:26.377 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:23
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.054)       0:00:26.431 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sdb"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:27
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.052)       0:00:26.484 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:34
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.049)       0:00:26.534 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:38
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.039)       0:00:26.573 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:42
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.050)       0:00:26.624 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:46
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.025)       0:00:26.650 ********* 
ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sdb"
}

MSG:

All assertions passed

TASK [Check MD RAID] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:56
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.042)       0:00:26.692 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-md.yml for /cache/rhel-7.qcow2

TASK [get information about RAID] **********************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:6
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.042)       0:00:26.735 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:12
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.022)       0:00:26.758 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:16
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.022)       0:00:26.780 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:20
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.024)       0:00:26.805 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:24
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.026)       0:00:26.832 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:30
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.023)       0:00:26.855 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:36
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.022)       0:00:26.877 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:44
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.022)       0:00:26.899 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:59
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.036)       0:00:26.936 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml:1
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.043)       0:00:26.979 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'})

TASK [Get information about LVM RAID] ******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.041)       0:00:27.021 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is LVM RAID] *******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.029)       0:00:27.050 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.028)       0:00:27.078 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:62
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.028)       0:00:27.107 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml for /cache/rhel-7.qcow2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml:1
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.046)       0:00:27.153 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'})

TASK [Get information about thinpool] ******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.043)       0:00:27.196 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8
Thursday 21 July 2022  15:43:28 +0000 (0:00:00.023)       0:00:27.219 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.024)       0:00:27.244 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.023)       0:00:27.267 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check member encryption] *************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:65
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.024)       0:00:27.292 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:4
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.047)       0:00:27.340 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:8
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.086)       0:00:27.427 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=/dev/sdb)  => {
    "_storage_test_pool_member_path": "/dev/sdb",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:15
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.027)       0:00:27.455 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/sdb)

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:1
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.072)       0:00:27.527 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:6
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.049)       0:00:27.576 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:11
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.048)       0:00:27.625 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:17
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.035)       0:00:27.660 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:23
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.036)       0:00:27.697 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:29
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.036)       0:00:27.733 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:22
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.034)       0:00:27.767 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:68
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.030)       0:00:27.798 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml:1
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.044)       0:00:27.842 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'})

TASK [get information about VDO deduplication] *********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.042)       0:00:27.884 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.021)       0:00:27.906 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.021)       0:00:27.927 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.021)       0:00:27.949 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.020)       0:00:27.970 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.021)       0:00:27.991 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.020)       0:00:28.012 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.022)       0:00:28.035 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:71
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.031)       0:00:28.066 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [verify the volumes] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml:3
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.036)       0:00:28.103 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'})

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.041)       0:00:28.145 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10
Thursday 21 July 2022  15:43:29 +0000 (0:00:00.045)       0:00:28.190 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache)

TASK [Get expected mount device based on device type] **************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6
Thursday 21 July 2022  15:43:30 +0000 (0:00:00.078)       0:00:28.269 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14
Thursday 21 July 2022  15:43:30 +0000 (0:00:00.039)       0:00:28.309 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": [
            {
                "block_available": 938726,
                "block_size": 4096,
                "block_total": 999344,
                "block_used": 60618,
                "device": "/dev/mapper/foo-test1",
                "fstype": "ext4",
                "inode_available": 262133,
                "inode_total": 262144,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 3845021696,
                "size_total": 4093313024,
                "uuid": "9ac0b734-2f82-4721-a26a-c5ea2621165e"
            }
        ],
        "storage_test_mount_expected_match_count": "1",
        "storage_test_mount_point_matches": [
            {
                "block_available": 938726,
                "block_size": 4096,
                "block_total": 999344,
                "block_used": 60618,
                "device": "/dev/mapper/foo-test1",
                "fstype": "ext4",
                "inode_available": 262133,
                "inode_total": 262144,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 3845021696,
                "size_total": 4093313024,
                "uuid": "9ac0b734-2f82-4721-a26a-c5ea2621165e"
            }
        ],
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28
Thursday 21 July 2022  15:43:30 +0000 (0:00:00.054)       0:00:28.363 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the current mount state by mount point] ***************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37
Thursday 21 July 2022  15:43:30 +0000 (0:00:00.050)       0:00:28.414 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the mount fs type] ************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45
Thursday 21 July 2022  15:43:30 +0000 (0:00:00.065)       0:00:28.479 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [command] *****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54
Thursday 21 July 2022  15:43:30 +0000 (0:00:00.050)       0:00:28.530 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58
Thursday 21 July 2022  15:43:30 +0000 (0:00:00.022)       0:00:28.553 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63
Thursday 21 July 2022  15:43:30 +0000 (0:00:00.022)       0:00:28.575 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75
Thursday 21 July 2022  15:43:30 +0000 (0:00:00.023)       0:00:28.598 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": null,
        "storage_test_mount_expected_match_count": null,
        "storage_test_mount_point_matches": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2
Thursday 21 July 2022  15:43:30 +0000 (0:00:00.032)       0:00:28.630 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25
Thursday 21 July 2022  15:43:30 +0000 (0:00:00.090)       0:00:28.721 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32
Thursday 21 July 2022  15:43:30 +0000 (0:00:00.112)       0:00:28.833 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39
Thursday 21 July 2022  15:43:30 +0000 (0:00:00.047)       0:00:28.881 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up variables] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49
Thursday 21 July 2022  15:43:30 +0000 (0:00:00.036)       0:00:28.917 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4
Thursday 21 July 2022  15:43:30 +0000 (0:00:00.036)       0:00:28.954 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10
Thursday 21 July 2022  15:43:30 +0000 (0:00:00.037)       0:00:28.992 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4
Thursday 21 July 2022  15:43:30 +0000 (0:00:00.036)       0:00:29.028 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658418202.7839499,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1658418202.7839499,
        "dev": 5,
        "device_type": 64512,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 31730,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1658418202.7839499,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10
Thursday 21 July 2022  15:43:31 +0000 (0:00:00.323)       0:00:29.352 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18
Thursday 21 July 2022  15:43:31 +0000 (0:00:00.038)       0:00:29.390 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [(1/2) Process volume type (set initial value)] ***************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24
Thursday 21 July 2022  15:43:31 +0000 (0:00:00.037)       0:00:29.428 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [(2/2) Process volume type (get RAID value)] ******************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28
Thursday 21 July 2022  15:43:31 +0000 (0:00:00.033)       0:00:29.461 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33
Thursday 21 July 2022  15:43:31 +0000 (0:00:00.021)       0:00:29.483 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3
Thursday 21 July 2022  15:43:31 +0000 (0:00:00.038)       0:00:29.521 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10
Thursday 21 July 2022  15:43:31 +0000 (0:00:00.024)       0:00:29.546 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "changed": true,
    "changes": {
        "installed": [
            "cryptsetup"
        ]
    },
    "rc": 0,
    "results": [
        "Loaded plugins: search-disabled-repos\nResolving Dependencies\n--> Running transaction check\n---> Package cryptsetup.x86_64 0:2.0.3-6.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package             Arch            Version                Repository     Size\n================================================================================\nInstalling:\n cryptsetup          x86_64          2.0.3-6.el7            rhel          154 k\n\nTransaction Summary\n================================================================================\nInstall  1 Package\n\nTotal download size: 154 k\nInstalled size: 354 k\nDownloading packages:\nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n  Installing : cryptsetup-2.0.3-6.el7.x86_64                                1/1 \n  Verifying  : cryptsetup-2.0.3-6.el7.x86_64                                1/1 \n\nInstalled:\n  cryptsetup.x86_64 0:2.0.3-6.el7                                               \n\nComplete!\n"
    ]
}

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15
Thursday 21 July 2022  15:43:32 +0000 (0:00:01.558)       0:00:31.105 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21
Thursday 21 July 2022  15:43:32 +0000 (0:00:00.026)       0:00:31.131 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30
Thursday 21 July 2022  15:43:32 +0000 (0:00:00.023)       0:00:31.155 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38
Thursday 21 July 2022  15:43:32 +0000 (0:00:00.049)       0:00:31.204 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44
Thursday 21 July 2022  15:43:32 +0000 (0:00:00.022)       0:00:31.227 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49
Thursday 21 July 2022  15:43:33 +0000 (0:00:00.022)       0:00:31.249 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55
Thursday 21 July 2022  15:43:33 +0000 (0:00:00.025)       0:00:31.275 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61
Thursday 21 July 2022  15:43:33 +0000 (0:00:00.025)       0:00:31.301 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67
Thursday 21 July 2022  15:43:33 +0000 (0:00:00.025)       0:00:31.326 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74
Thursday 21 July 2022  15:43:33 +0000 (0:00:00.081)       0:00:31.408 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79
Thursday 21 July 2022  15:43:33 +0000 (0:00:00.052)       0:00:31.460 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85
Thursday 21 July 2022  15:43:33 +0000 (0:00:00.037)       0:00:31.498 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91
Thursday 21 July 2022  15:43:33 +0000 (0:00:00.037)       0:00:31.535 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97
Thursday 21 July 2022  15:43:33 +0000 (0:00:00.038)       0:00:31.574 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [get information about RAID] **********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7
Thursday 21 July 2022  15:43:33 +0000 (0:00:00.071)       0:00:31.646 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13
Thursday 21 July 2022  15:43:33 +0000 (0:00:00.038)       0:00:31.685 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17
Thursday 21 July 2022  15:43:33 +0000 (0:00:00.036)       0:00:31.722 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21
Thursday 21 July 2022  15:43:33 +0000 (0:00:00.039)       0:00:31.761 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25
Thursday 21 July 2022  15:43:33 +0000 (0:00:00.037)       0:00:31.799 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31
Thursday 21 July 2022  15:43:33 +0000 (0:00:00.038)       0:00:31.837 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37
Thursday 21 July 2022  15:43:33 +0000 (0:00:00.042)       0:00:31.880 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the actual size of the volume] *************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3
Thursday 21 July 2022  15:43:33 +0000 (0:00:00.042)       0:00:31.922 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [parse the requested size of the volume] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9
Thursday 21 July 2022  15:43:34 +0000 (0:00:00.413)       0:00:32.335 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15
Thursday 21 July 2022  15:43:34 +0000 (0:00:00.332)       0:00:32.667 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_expected_size": "4294967296"
    },
    "changed": false
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20
Thursday 21 July 2022  15:43:34 +0000 (0:00:00.050)       0:00:32.718 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25
Thursday 21 July 2022  15:43:34 +0000 (0:00:00.039)       0:00:32.758 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28
Thursday 21 July 2022  15:43:34 +0000 (0:00:00.041)       0:00:32.799 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31
Thursday 21 July 2022  15:43:34 +0000 (0:00:00.037)       0:00:32.837 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36
Thursday 21 July 2022  15:43:34 +0000 (0:00:00.041)       0:00:32.878 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39
Thursday 21 July 2022  15:43:34 +0000 (0:00:00.036)       0:00:32.914 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44
Thursday 21 July 2022  15:43:34 +0000 (0:00:00.034)       0:00:32.948 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_actual_size": {
        "bytes": 4294967296,
        "changed": false,
        "failed": false,
        "lvm": "4g",
        "parted": "4GiB",
        "size": "4 GiB"
    }
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47
Thursday 21 July 2022  15:43:34 +0000 (0:00:00.033)       0:00:32.982 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50
Thursday 21 July 2022  15:43:34 +0000 (0:00:00.033)       0:00:33.016 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6
Thursday 21 July 2022  15:43:34 +0000 (0:00:00.047)       0:00:33.063 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.025214",
    "end": "2022-07-21 11:43:34.975502",
    "rc": 0,
    "start": "2022-07-21 11:43:34.950288"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14
Thursday 21 July 2022  15:43:35 +0000 (0:00:00.340)       0:00:33.403 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [check segment type] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17
Thursday 21 July 2022  15:43:35 +0000 (0:00:00.053)       0:00:33.457 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22
Thursday 21 July 2022  15:43:35 +0000 (0:00:00.064)       0:00:33.521 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested cache size] ******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26
Thursday 21 July 2022  15:43:35 +0000 (0:00:00.043)       0:00:33.564 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32
Thursday 21 July 2022  15:43:35 +0000 (0:00:00.042)       0:00:33.607 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36
Thursday 21 July 2022  15:43:35 +0000 (0:00:00.044)       0:00:33.652 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16
Thursday 21 July 2022  15:43:35 +0000 (0:00:00.039)       0:00:33.691 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43
Thursday 21 July 2022  15:43:35 +0000 (0:00:00.086)       0:00:33.777 ********* 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53
Thursday 21 July 2022  15:43:35 +0000 (0:00:00.021)       0:00:33.799 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Remove the volume group created above] ***********************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:48
Thursday 21 July 2022  15:43:35 +0000 (0:00:00.077)       0:00:33.877 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:43:35 +0000 (0:00:00.046)       0:00:33.923 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:43:35 +0000 (0:00:00.031)       0:00:33.955 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:43:36 +0000 (0:00:00.511)       0:00:34.467 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:43:36 +0000 (0:00:00.058)       0:00:34.526 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:43:36 +0000 (0:00:00.031)       0:00:34.557 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:43:36 +0000 (0:00:00.031)       0:00:34.589 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:43:36 +0000 (0:00:00.044)       0:00:34.634 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:43:36 +0000 (0:00:00.021)       0:00:34.655 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:43:36 +0000 (0:00:00.029)       0:00:34.684 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "sdb"
            ],
            "name": "foo",
            "state": "absent",
            "type": "lvm"
        }
    ]
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:43:36 +0000 (0:00:00.039)       0:00:34.724 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:43:36 +0000 (0:00:00.036)       0:00:34.760 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:43:36 +0000 (0:00:00.029)       0:00:34.789 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:43:36 +0000 (0:00:00.027)       0:00:34.817 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:43:36 +0000 (0:00:00.030)       0:00:34.848 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:43:36 +0000 (0:00:00.029)       0:00:34.877 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:43:36 +0000 (0:00:00.060)       0:00:34.938 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:43:36 +0000 (0:00:00.023)       0:00:34.962 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "destroy device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/sdb",
            "fs_type": "lvmpv"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [
        {
            "fstype": "ext4",
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "absent"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "absent",
            "type": "lvm",
            "volumes": []
        }
    ],
    "volumes": []
}

TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91
Thursday 21 July 2022  15:43:38 +0000 (0:00:01.594)       0:00:36.557 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:43:38 +0000 (0:00:00.037)       0:00:36.594 ********* 

TASK [linux-system-roles.storage : show blivet_output] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109
Thursday 21 July 2022  15:43:38 +0000 (0:00:00.023)       0:00:36.617 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "destroy device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/sdb",
                "fs_type": "lvmpv"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [
            {
                "fstype": "ext4",
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "absent"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ],
        "volumes": []
    }
}

TASK [linux-system-roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114
Thursday 21 July 2022  15:43:38 +0000 (0:00:00.037)       0:00:36.655 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ]
    },
    "changed": false
}

TASK [linux-system-roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118
Thursday 21 July 2022  15:43:38 +0000 (0:00:00.069)       0:00:36.725 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : remove obsolete mounts] *********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134
Thursday 21 July 2022  15:43:38 +0000 (0:00:00.070)       0:00:36.795 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'ext4'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "fstype": "ext4",
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146
Thursday 21 July 2022  15:43:38 +0000 (0:00:00.376)       0:00:37.171 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : set up new/current mounts] ******************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151
Thursday 21 July 2022  15:43:39 +0000 (0:00:00.471)       0:00:37.643 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163
Thursday 21 July 2022  15:43:39 +0000 (0:00:00.039)       0:00:37.682 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171
Thursday 21 July 2022  15:43:39 +0000 (0:00:00.472)       0:00:38.154 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658418192.4109497,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176
Thursday 21 July 2022  15:43:40 +0000 (0:00:00.307)       0:00:38.461 ********* 

TASK [linux-system-roles.storage : Update facts] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198
Thursday 21 July 2022  15:43:40 +0000 (0:00:00.023)       0:00:38.485 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:58
Thursday 21 July 2022  15:43:41 +0000 (0:00:00.850)       0:00:39.335 ********* 
included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1
Thursday 21 July 2022  15:43:41 +0000 (0:00:00.043)       0:00:39.378 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "absent",
            "type": "lvm",
            "volumes": []
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6
Thursday 21 July 2022  15:43:41 +0000 (0:00:00.050)       0:00:39.429 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14
Thursday 21 July 2022  15:43:41 +0000 (0:00:00.035)       0:00:39.465 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-15-42-49-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19
Thursday 21 July 2022  15:43:41 +0000 (0:00:00.352)       0:00:39.817 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003610",
    "end": "2022-07-21 11:43:41.701169",
    "rc": 0,
    "start": "2022-07-21 11:43:41.697559"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24
Thursday 21 July 2022  15:43:41 +0000 (0:00:00.311)       0:00:40.128 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003453",
    "end": "2022-07-21 11:43:42.016684",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 11:43:42.013231"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.318)       0:00:40.447 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'absent', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None})

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:5
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.056)       0:00:40.504 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:18
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.072)       0:00:40.576 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members)
included: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes)

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:1
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.074)       0:00:40.651 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "0",
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:10
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.056)       0:00:40.707 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:19
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.020)       0:00:40.728 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": "0"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:23
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.049)       0:00:40.778 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": []
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:27
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.050)       0:00:40.828 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:34
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.051)       0:00:40.879 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:38
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.034)       0:00:40.914 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:42
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.050)       0:00:40.965 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:46
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.024)       0:00:40.989 ********* 

TASK [Check MD RAID] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:56
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.019)       0:00:41.009 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-md.yml for /cache/rhel-7.qcow2

TASK [get information about RAID] **********************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:6
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.039)       0:00:41.049 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:12
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.022)       0:00:41.072 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:16
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.022)       0:00:41.094 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:20
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.022)       0:00:41.116 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:24
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.021)       0:00:41.138 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:30
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.027)       0:00:41.165 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:36
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.025)       0:00:41.190 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:44
Thursday 21 July 2022  15:43:42 +0000 (0:00:00.025)       0:00:41.215 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:59
Thursday 21 July 2022  15:43:43 +0000 (0:00:00.035)       0:00:41.251 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml:1
Thursday 21 July 2022  15:43:43 +0000 (0:00:00.046)       0:00:41.297 ********* 

TASK [Check Thin Pools] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:62
Thursday 21 July 2022  15:43:43 +0000 (0:00:00.021)       0:00:41.319 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml for /cache/rhel-7.qcow2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml:1
Thursday 21 July 2022  15:43:43 +0000 (0:00:00.042)       0:00:41.362 ********* 

TASK [Check member encryption] *************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:65
Thursday 21 July 2022  15:43:43 +0000 (0:00:00.020)       0:00:41.382 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:4
Thursday 21 July 2022  15:43:43 +0000 (0:00:00.043)       0:00:41.425 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:8
Thursday 21 July 2022  15:43:43 +0000 (0:00:00.047)       0:00:41.472 ********* 

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:15
Thursday 21 July 2022  15:43:43 +0000 (0:00:00.021)       0:00:41.494 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:22
Thursday 21 July 2022  15:43:43 +0000 (0:00:00.019)       0:00:41.514 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:68
Thursday 21 July 2022  15:43:43 +0000 (0:00:00.034)       0:00:41.548 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml:1
Thursday 21 July 2022  15:43:43 +0000 (0:00:00.045)       0:00:41.593 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:71
Thursday 21 July 2022  15:43:43 +0000 (0:00:00.021)       0:00:41.615 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [verify the volumes] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml:3
Thursday 21 July 2022  15:43:43 +0000 (0:00:00.041)       0:00:41.656 ********* 

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43
Thursday 21 July 2022  15:43:43 +0000 (0:00:00.020)       0:00:41.677 ********* 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53
Thursday 21 July 2022  15:43:43 +0000 (0:00:00.021)       0:00:41.699 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Try to create ext4 filesystem with invalid parameter "-Fb 512"] **********
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:62
Thursday 21 July 2022  15:43:43 +0000 (0:00:00.076)       0:00:41.775 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:43:43 +0000 (0:00:00.072)       0:00:41.847 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:43:43 +0000 (0:00:00.034)       0:00:41.882 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:43:44 +0000 (0:00:00.419)       0:00:42.301 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:43:44 +0000 (0:00:00.061)       0:00:42.363 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:43:44 +0000 (0:00:00.034)       0:00:42.397 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:43:44 +0000 (0:00:00.033)       0:00:42.431 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:43:44 +0000 (0:00:00.046)       0:00:42.477 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:43:44 +0000 (0:00:00.021)       0:00:42.499 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:43:44 +0000 (0:00:00.027)       0:00:42.527 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "sdb"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_create_options": "-Fb 512",
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:43:44 +0000 (0:00:00.039)       0:00:42.566 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:43:44 +0000 (0:00:00.037)       0:00:42.604 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:43:44 +0000 (0:00:00.033)       0:00:42.637 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:43:44 +0000 (0:00:00.028)       0:00:42.666 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:43:44 +0000 (0:00:00.026)       0:00:42.692 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:43:44 +0000 (0:00:00.025)       0:00:42.717 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:43:44 +0000 (0:00:00.052)       0:00:42.770 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:43:44 +0000 (0:00:00.020)       0:00:42.791 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "xfsprogs",
        "e2fsprogs",
        "lvm2"
    ],
    "pools": [],
    "volumes": []
}

MSG:

Failed to commit changes to disk: (FSError('format failed: 1',), '/dev/mapper/foo-test1')

TASK [linux-system-roles.storage : failed message] *****************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99
Thursday 21 July 2022  15:43:45 +0000 (0:00:01.334)       0:00:44.126 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "changed": false
}

MSG:

{'crypts': [], 'mounts': [], 'leaves': [], 'changed': False, 'actions': [], 'failed': True, 'volumes': [], 'invocation': {'module_args': {'packages_only': False, 'disklabel_type': None, 'diskvolume_mkfs_option_map': {'ext4': '-F', 'ext3': '-F', 'ext2': '-F'}, 'safe_mode': False, 'pools': [{'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 512'}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}], 'volumes': [], 'pool_defaults': {'encryption_password': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_cipher': None, 'disks': [], 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_device_count': None, 'state': 'present', 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'raid_level': None, 'raid_spare_count': None}, 'volume_defaults': {'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'xfs', 'mount_options': 'defaults', 'size': 0, 'mount_point': '', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'fs_overwrite_existing': True, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'mount_passno': 0, 'raid_spare_count': None, 'cache_mode': None, 'deduplication': None, 'cached': False, 'type': 'lvm', 'disks': [], 'thin_pool_size': None, 'thin': None, 'mount_check': 0, 'cache_size': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, 'use_partitions': None}}, 'pools': [], 'packages': ['xfsprogs', 'e2fsprogs', 'lvm2'], 'msg': "Failed to commit changes to disk: (FSError('format failed: 1',), '/dev/mapper/foo-test1')", '_ansible_no_log': False}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:43:45 +0000 (0:00:00.039)       0:00:44.165 ********* 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:82
Thursday 21 July 2022  15:43:45 +0000 (0:00:00.022)       0:00:44.187 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the output when creating ext4 filesystem with invalid parameter "-Fb 512"] ***
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:88
Thursday 21 July 2022  15:43:45 +0000 (0:00:00.038)       0:00:44.226 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Remove the volume group created above] ***********************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:95
Thursday 21 July 2022  15:43:46 +0000 (0:00:00.059)       0:00:44.286 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:43:46 +0000 (0:00:00.036)       0:00:44.323 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:43:46 +0000 (0:00:00.034)       0:00:44.357 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:43:46 +0000 (0:00:00.430)       0:00:44.787 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:43:46 +0000 (0:00:00.148)       0:00:44.936 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:43:46 +0000 (0:00:00.033)       0:00:44.970 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:43:46 +0000 (0:00:00.033)       0:00:45.003 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:43:46 +0000 (0:00:00.046)       0:00:45.049 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:43:46 +0000 (0:00:00.019)       0:00:45.068 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:43:46 +0000 (0:00:00.026)       0:00:45.095 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "sdb"
            ],
            "name": "foo",
            "state": "absent",
            "type": "lvm"
        }
    ]
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:43:46 +0000 (0:00:00.035)       0:00:45.131 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:43:46 +0000 (0:00:00.035)       0:00:45.166 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:43:46 +0000 (0:00:00.027)       0:00:45.193 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:43:46 +0000 (0:00:00.032)       0:00:45.225 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:43:47 +0000 (0:00:00.029)       0:00:45.255 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:43:47 +0000 (0:00:00.032)       0:00:45.288 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:43:47 +0000 (0:00:00.057)       0:00:45.346 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:43:47 +0000 (0:00:00.020)       0:00:45.366 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "destroy device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/sdb",
            "fs_type": "lvmpv"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [],
    "packages": [
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "absent",
            "type": "lvm",
            "volumes": []
        }
    ],
    "volumes": []
}

TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91
Thursday 21 July 2022  15:43:48 +0000 (0:00:01.358)       0:00:46.725 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:43:48 +0000 (0:00:00.036)       0:00:46.761 ********* 

TASK [linux-system-roles.storage : show blivet_output] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109
Thursday 21 July 2022  15:43:48 +0000 (0:00:00.022)       0:00:46.784 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "destroy device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/sdb",
                "fs_type": "lvmpv"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [],
        "packages": [
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ],
        "volumes": []
    }
}

TASK [linux-system-roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114
Thursday 21 July 2022  15:43:48 +0000 (0:00:00.040)       0:00:46.824 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ]
    },
    "changed": false
}

TASK [linux-system-roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118
Thursday 21 July 2022  15:43:48 +0000 (0:00:00.034)       0:00:46.859 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : remove obsolete mounts] *********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134
Thursday 21 July 2022  15:43:48 +0000 (0:00:00.036)       0:00:46.895 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146
Thursday 21 July 2022  15:43:48 +0000 (0:00:00.035)       0:00:46.930 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : set up new/current mounts] ******************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151
Thursday 21 July 2022  15:43:48 +0000 (0:00:00.027)       0:00:46.957 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163
Thursday 21 July 2022  15:43:48 +0000 (0:00:00.037)       0:00:46.994 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171
Thursday 21 July 2022  15:43:48 +0000 (0:00:00.022)       0:00:47.017 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658418192.4109497,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176
Thursday 21 July 2022  15:43:49 +0000 (0:00:00.309)       0:00:47.326 ********* 

TASK [linux-system-roles.storage : Update facts] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198
Thursday 21 July 2022  15:43:49 +0000 (0:00:00.022)       0:00:47.349 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [Create one LVM logical volume with "4g" under one volume group] **********
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:105
Thursday 21 July 2022  15:43:49 +0000 (0:00:00.829)       0:00:48.179 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:43:49 +0000 (0:00:00.040)       0:00:48.219 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:43:50 +0000 (0:00:00.034)       0:00:48.253 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:43:50 +0000 (0:00:00.419)       0:00:48.673 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:43:50 +0000 (0:00:00.099)       0:00:48.772 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:43:50 +0000 (0:00:00.099)       0:00:48.872 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:43:50 +0000 (0:00:00.042)       0:00:48.914 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:43:50 +0000 (0:00:00.048)       0:00:48.963 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:43:50 +0000 (0:00:00.022)       0:00:48.985 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:43:50 +0000 (0:00:00.031)       0:00:49.017 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "sdb"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:43:50 +0000 (0:00:00.043)       0:00:49.060 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:43:50 +0000 (0:00:00.040)       0:00:49.101 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:43:50 +0000 (0:00:00.028)       0:00:49.129 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:43:50 +0000 (0:00:00.025)       0:00:49.155 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:43:50 +0000 (0:00:00.024)       0:00:49.180 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:43:50 +0000 (0:00:00.032)       0:00:49.213 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:43:51 +0000 (0:00:00.053)       0:00:49.266 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:43:51 +0000 (0:00:00.019)       0:00:49.286 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "create format",
            "device": "/dev/sdb",
            "fs_type": "lvmpv"
        },
        {
            "action": "create device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "create device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd",
        "/dev/mapper/foo-test1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext4",
            "opts": "defaults",
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs",
        "lvm2"
    ],
    "pools": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91
Thursday 21 July 2022  15:43:52 +0000 (0:00:01.622)       0:00:50.908 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:43:52 +0000 (0:00:00.035)       0:00:50.943 ********* 

TASK [linux-system-roles.storage : show blivet_output] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109
Thursday 21 July 2022  15:43:52 +0000 (0:00:00.020)       0:00:50.963 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "create format",
                "device": "/dev/sdb",
                "fs_type": "lvmpv"
            },
            {
                "action": "create device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "create device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd",
            "/dev/mapper/foo-test1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext4",
                "opts": "defaults",
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs",
            "lvm2"
        ],
        "pools": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [linux-system-roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114
Thursday 21 July 2022  15:43:52 +0000 (0:00:00.037)       0:00:51.001 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [linux-system-roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118
Thursday 21 July 2022  15:43:52 +0000 (0:00:00.033)       0:00:51.035 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : remove obsolete mounts] *********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134
Thursday 21 July 2022  15:43:52 +0000 (0:00:00.034)       0:00:51.070 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146
Thursday 21 July 2022  15:43:52 +0000 (0:00:00.035)       0:00:51.105 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : set up new/current mounts] ******************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151
Thursday 21 July 2022  15:43:53 +0000 (0:00:00.475)       0:00:51.581 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'ext4', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "opts": "defaults",
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163
Thursday 21 July 2022  15:43:53 +0000 (0:00:00.365)       0:00:51.946 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171
Thursday 21 July 2022  15:43:54 +0000 (0:00:00.485)       0:00:52.432 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658418192.4109497,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176
Thursday 21 July 2022  15:43:54 +0000 (0:00:00.325)       0:00:52.758 ********* 

TASK [linux-system-roles.storage : Update facts] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198
Thursday 21 July 2022  15:43:54 +0000 (0:00:00.021)       0:00:52.779 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:119
Thursday 21 July 2022  15:43:55 +0000 (0:00:00.857)       0:00:53.637 ********* 
included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1
Thursday 21 July 2022  15:43:55 +0000 (0:00:00.039)       0:00:53.676 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6
Thursday 21 July 2022  15:43:55 +0000 (0:00:00.121)       0:00:53.798 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14
Thursday 21 July 2022  15:43:55 +0000 (0:00:00.044)       0:00:53.843 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/mapper/foo-test1": {
            "fstype": "ext4",
            "label": "",
            "name": "/dev/mapper/foo-test1",
            "size": "4G",
            "type": "lvm",
            "uuid": "561ab946-7428-4dac-ada8-aef5f1427d66"
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "LVM2_member",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": "tznhVH-nMhW-rvxB-06Ta-lvZF-QQuu-PAfX1t"
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-15-42-49-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19
Thursday 21 July 2022  15:43:56 +0000 (0:00:01.317)       0:00:55.160 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003612",
    "end": "2022-07-21 11:43:57.046280",
    "rc": 0,
    "start": "2022-07-21 11:43:57.042668"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0
/dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24
Thursday 21 July 2022  15:43:57 +0000 (0:00:00.315)       0:00:55.476 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003593",
    "end": "2022-07-21 11:43:57.362580",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 11:43:57.358987"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33
Thursday 21 July 2022  15:43:57 +0000 (0:00:00.316)       0:00:55.792 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None})

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:5
Thursday 21 July 2022  15:43:57 +0000 (0:00:00.057)       0:00:55.850 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:18
Thursday 21 July 2022  15:43:57 +0000 (0:00:00.032)       0:00:55.882 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members)
included: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes)

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:1
Thursday 21 July 2022  15:43:57 +0000 (0:00:00.045)       0:00:55.927 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sdb"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:10
Thursday 21 July 2022  15:43:57 +0000 (0:00:00.056)       0:00:55.984 ********* 
ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sdb",
    "pv": "/dev/sdb"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:19
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.306)       0:00:56.291 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:23
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.046)       0:00:56.337 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sdb"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:27
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.046)       0:00:56.384 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:34
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.047)       0:00:56.432 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:38
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.036)       0:00:56.469 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:42
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.045)       0:00:56.514 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:46
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.022)       0:00:56.537 ********* 
ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sdb"
}

MSG:

All assertions passed

TASK [Check MD RAID] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:56
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.041)       0:00:56.578 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-md.yml for /cache/rhel-7.qcow2

TASK [get information about RAID] **********************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:6
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.043)       0:00:56.622 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:12
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.023)       0:00:56.645 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:16
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.022)       0:00:56.668 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:20
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.021)       0:00:56.690 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:24
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.021)       0:00:56.711 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:30
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.021)       0:00:56.732 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:36
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.022)       0:00:56.754 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:44
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.052)       0:00:56.806 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:59
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.032)       0:00:56.839 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml:1
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.041)       0:00:56.880 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [Get information about LVM RAID] ******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.039)       0:00:56.920 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is LVM RAID] *******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.024)       0:00:56.945 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.025)       0:00:56.970 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:62
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.024)       0:00:56.995 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml for /cache/rhel-7.qcow2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml:1
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.046)       0:00:57.042 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [Get information about thinpool] ******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.041)       0:00:57.083 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.021)       0:00:57.105 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.022)       0:00:57.127 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.023)       0:00:57.151 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check member encryption] *************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:65
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.025)       0:00:57.177 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:4
Thursday 21 July 2022  15:43:58 +0000 (0:00:00.048)       0:00:57.226 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:8
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.053)       0:00:57.279 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=/dev/sdb)  => {
    "_storage_test_pool_member_path": "/dev/sdb",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:15
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.026)       0:00:57.305 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/sdb)

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:1
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.041)       0:00:57.347 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:6
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.053)       0:00:57.400 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:11
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.057)       0:00:57.457 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:17
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.035)       0:00:57.493 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:23
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.037)       0:00:57.530 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:29
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.034)       0:00:57.565 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:22
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.033)       0:00:57.599 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:68
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.035)       0:00:57.634 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml:1
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.051)       0:00:57.685 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [get information about VDO deduplication] *********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.046)       0:00:57.732 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.024)       0:00:57.756 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.022)       0:00:57.779 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.022)       0:00:57.802 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.022)       0:00:57.824 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.023)       0:00:57.848 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.023)       0:00:57.871 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.025)       0:00:57.896 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:71
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.079)       0:00:57.975 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [verify the volumes] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml:3
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.100)       0:00:58.076 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.040)       0:00:58.117 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10
Thursday 21 July 2022  15:43:59 +0000 (0:00:00.052)       0:00:58.169 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache)

TASK [Get expected mount device based on device type] **************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6
Thursday 21 July 2022  15:44:00 +0000 (0:00:00.079)       0:00:58.248 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14
Thursday 21 July 2022  15:44:00 +0000 (0:00:00.042)       0:00:58.290 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": [
            {
                "block_available": 938726,
                "block_size": 4096,
                "block_total": 999344,
                "block_used": 60618,
                "device": "/dev/mapper/foo-test1",
                "fstype": "ext4",
                "inode_available": 262133,
                "inode_total": 262144,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 3845021696,
                "size_total": 4093313024,
                "uuid": "561ab946-7428-4dac-ada8-aef5f1427d66"
            }
        ],
        "storage_test_mount_expected_match_count": "1",
        "storage_test_mount_point_matches": [
            {
                "block_available": 938726,
                "block_size": 4096,
                "block_total": 999344,
                "block_used": 60618,
                "device": "/dev/mapper/foo-test1",
                "fstype": "ext4",
                "inode_available": 262133,
                "inode_total": 262144,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 3845021696,
                "size_total": 4093313024,
                "uuid": "561ab946-7428-4dac-ada8-aef5f1427d66"
            }
        ],
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28
Thursday 21 July 2022  15:44:00 +0000 (0:00:00.057)       0:00:58.348 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the current mount state by mount point] ***************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37
Thursday 21 July 2022  15:44:00 +0000 (0:00:00.062)       0:00:58.410 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the mount fs type] ************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45
Thursday 21 July 2022  15:44:00 +0000 (0:00:00.050)       0:00:58.460 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [command] *****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54
Thursday 21 July 2022  15:44:00 +0000 (0:00:00.051)       0:00:58.511 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58
Thursday 21 July 2022  15:44:00 +0000 (0:00:00.022)       0:00:58.534 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63
Thursday 21 July 2022  15:44:00 +0000 (0:00:00.023)       0:00:58.558 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75
Thursday 21 July 2022  15:44:00 +0000 (0:00:00.022)       0:00:58.580 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": null,
        "storage_test_mount_expected_match_count": null,
        "storage_test_mount_point_matches": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2
Thursday 21 July 2022  15:44:00 +0000 (0:00:00.034)       0:00:58.615 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25
Thursday 21 July 2022  15:44:00 +0000 (0:00:00.059)       0:00:58.674 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32
Thursday 21 July 2022  15:44:00 +0000 (0:00:00.050)       0:00:58.725 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39
Thursday 21 July 2022  15:44:00 +0000 (0:00:00.048)       0:00:58.773 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up variables] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49
Thursday 21 July 2022  15:44:00 +0000 (0:00:00.034)       0:00:58.808 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4
Thursday 21 July 2022  15:44:00 +0000 (0:00:00.037)       0:00:58.846 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10
Thursday 21 July 2022  15:44:00 +0000 (0:00:00.039)       0:00:58.885 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4
Thursday 21 July 2022  15:44:00 +0000 (0:00:00.040)       0:00:58.925 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658418232.4419498,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1658418232.4419498,
        "dev": 5,
        "device_type": 64512,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 42550,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1658418232.4419498,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10
Thursday 21 July 2022  15:44:01 +0000 (0:00:00.319)       0:00:59.245 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18
Thursday 21 July 2022  15:44:01 +0000 (0:00:00.040)       0:00:59.285 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [(1/2) Process volume type (set initial value)] ***************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24
Thursday 21 July 2022  15:44:01 +0000 (0:00:00.039)       0:00:59.324 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [(2/2) Process volume type (get RAID value)] ******************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28
Thursday 21 July 2022  15:44:01 +0000 (0:00:00.035)       0:00:59.360 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33
Thursday 21 July 2022  15:44:01 +0000 (0:00:00.024)       0:00:59.384 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3
Thursday 21 July 2022  15:44:01 +0000 (0:00:00.038)       0:00:59.423 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10
Thursday 21 July 2022  15:44:01 +0000 (0:00:00.022)       0:00:59.445 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "rc": 0,
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15
Thursday 21 July 2022  15:44:01 +0000 (0:00:00.603)       0:01:00.048 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21
Thursday 21 July 2022  15:44:01 +0000 (0:00:00.022)       0:01:00.071 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30
Thursday 21 July 2022  15:44:01 +0000 (0:00:00.021)       0:01:00.093 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38
Thursday 21 July 2022  15:44:01 +0000 (0:00:00.048)       0:01:00.142 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44
Thursday 21 July 2022  15:44:01 +0000 (0:00:00.021)       0:01:00.163 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49
Thursday 21 July 2022  15:44:01 +0000 (0:00:00.021)       0:01:00.184 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55
Thursday 21 July 2022  15:44:01 +0000 (0:00:00.021)       0:01:00.205 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61
Thursday 21 July 2022  15:44:01 +0000 (0:00:00.022)       0:01:00.228 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67
Thursday 21 July 2022  15:44:02 +0000 (0:00:00.020)       0:01:00.249 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74
Thursday 21 July 2022  15:44:02 +0000 (0:00:00.047)       0:01:00.296 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79
Thursday 21 July 2022  15:44:02 +0000 (0:00:00.045)       0:01:00.342 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85
Thursday 21 July 2022  15:44:02 +0000 (0:00:00.036)       0:01:00.378 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91
Thursday 21 July 2022  15:44:02 +0000 (0:00:00.033)       0:01:00.411 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97
Thursday 21 July 2022  15:44:02 +0000 (0:00:00.032)       0:01:00.443 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [get information about RAID] **********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7
Thursday 21 July 2022  15:44:02 +0000 (0:00:00.032)       0:01:00.476 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13
Thursday 21 July 2022  15:44:02 +0000 (0:00:00.036)       0:01:00.512 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17
Thursday 21 July 2022  15:44:02 +0000 (0:00:00.033)       0:01:00.545 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21
Thursday 21 July 2022  15:44:02 +0000 (0:00:00.033)       0:01:00.578 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25
Thursday 21 July 2022  15:44:02 +0000 (0:00:00.032)       0:01:00.611 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31
Thursday 21 July 2022  15:44:02 +0000 (0:00:00.036)       0:01:00.647 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37
Thursday 21 July 2022  15:44:02 +0000 (0:00:00.033)       0:01:00.680 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the actual size of the volume] *************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3
Thursday 21 July 2022  15:44:02 +0000 (0:00:00.035)       0:01:00.715 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [parse the requested size of the volume] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9
Thursday 21 July 2022  15:44:02 +0000 (0:00:00.313)       0:01:01.029 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15
Thursday 21 July 2022  15:44:03 +0000 (0:00:00.329)       0:01:01.358 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_expected_size": "4294967296"
    },
    "changed": false
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20
Thursday 21 July 2022  15:44:03 +0000 (0:00:00.049)       0:01:01.407 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25
Thursday 21 July 2022  15:44:03 +0000 (0:00:00.040)       0:01:01.447 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28
Thursday 21 July 2022  15:44:03 +0000 (0:00:00.036)       0:01:01.484 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31
Thursday 21 July 2022  15:44:03 +0000 (0:00:00.039)       0:01:01.523 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36
Thursday 21 July 2022  15:44:03 +0000 (0:00:00.036)       0:01:01.560 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39
Thursday 21 July 2022  15:44:03 +0000 (0:00:00.035)       0:01:01.596 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44
Thursday 21 July 2022  15:44:03 +0000 (0:00:00.035)       0:01:01.631 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_actual_size": {
        "bytes": 4294967296,
        "changed": false,
        "failed": false,
        "lvm": "4g",
        "parted": "4GiB",
        "size": "4 GiB"
    }
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47
Thursday 21 July 2022  15:44:03 +0000 (0:00:00.036)       0:01:01.668 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50
Thursday 21 July 2022  15:44:03 +0000 (0:00:00.033)       0:01:01.701 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6
Thursday 21 July 2022  15:44:03 +0000 (0:00:00.079)       0:01:01.781 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.031294",
    "end": "2022-07-21 11:44:03.699429",
    "rc": 0,
    "start": "2022-07-21 11:44:03.668135"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14
Thursday 21 July 2022  15:44:03 +0000 (0:00:00.350)       0:01:02.131 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [check segment type] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17
Thursday 21 July 2022  15:44:03 +0000 (0:00:00.051)       0:01:02.183 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22
Thursday 21 July 2022  15:44:04 +0000 (0:00:00.082)       0:01:02.265 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested cache size] ******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26
Thursday 21 July 2022  15:44:04 +0000 (0:00:00.038)       0:01:02.303 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32
Thursday 21 July 2022  15:44:04 +0000 (0:00:00.079)       0:01:02.383 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36
Thursday 21 July 2022  15:44:04 +0000 (0:00:00.037)       0:01:02.420 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16
Thursday 21 July 2022  15:44:04 +0000 (0:00:00.039)       0:01:02.459 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43
Thursday 21 July 2022  15:44:04 +0000 (0:00:00.035)       0:01:02.495 ********* 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53
Thursday 21 July 2022  15:44:04 +0000 (0:00:00.020)       0:01:02.516 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Resizing with one large value which large than disk's size] **************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:123
Thursday 21 July 2022  15:44:04 +0000 (0:00:00.033)       0:01:02.549 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:44:04 +0000 (0:00:00.041)       0:01:02.591 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:44:04 +0000 (0:00:00.034)       0:01:02.625 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:44:04 +0000 (0:00:00.435)       0:01:03.060 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:44:04 +0000 (0:00:00.067)       0:01:03.128 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:44:04 +0000 (0:00:00.038)       0:01:03.167 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:44:04 +0000 (0:00:00.035)       0:01:03.202 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:44:05 +0000 (0:00:00.045)       0:01:03.247 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:44:05 +0000 (0:00:00.018)       0:01:03.266 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:44:05 +0000 (0:00:00.026)       0:01:03.292 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "sdb"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "12884901888.0"
                }
            ]
        }
    ]
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:44:05 +0000 (0:00:00.048)       0:01:03.341 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:44:05 +0000 (0:00:00.033)       0:01:03.375 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:44:05 +0000 (0:00:00.029)       0:01:03.404 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:44:05 +0000 (0:00:00.027)       0:01:03.432 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:44:05 +0000 (0:00:00.025)       0:01:03.458 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:44:05 +0000 (0:00:00.025)       0:01:03.483 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:44:05 +0000 (0:00:00.055)       0:01:03.538 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:44:05 +0000 (0:00:00.020)       0:01:03.559 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

volume 'test1' cannot be resized to '12 GiB'

TASK [linux-system-roles.storage : failed message] *****************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99
Thursday 21 July 2022  15:44:06 +0000 (0:00:01.307)       0:01:04.866 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "changed": false
}

MSG:

{'crypts': [], 'mounts': [], 'leaves': [], 'changed': False, 'actions': [], 'failed': True, 'volumes': [], 'invocation': {'module_args': {'packages_only': False, 'disklabel_type': None, 'diskvolume_mkfs_option_map': {'ext4': '-F', 'ext3': '-F', 'ext2': '-F'}, 'safe_mode': False, 'pools': [{'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', 'size': '12884901888.0', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}], 'volumes': [], 'pool_defaults': {'encryption_password': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_cipher': None, 'disks': [], 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_device_count': None, 'state': 'present', 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'raid_level': None, 'raid_spare_count': None}, 'volume_defaults': {'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'xfs', 'mount_options': 'defaults', 'size': 0, 'mount_point': '', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'fs_overwrite_existing': True, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'mount_passno': 0, 'raid_spare_count': None, 'cache_mode': None, 'deduplication': None, 'cached': False, 'type': 'lvm', 'disks': [], 'thin_pool_size': None, 'thin': None, 'mount_check': 0, 'cache_size': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, 'use_partitions': None}}, 'pools': [], 'packages': [], 'msg': "volume 'test1' cannot be resized to '12 GiB'", '_ansible_no_log': False}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:44:06 +0000 (0:00:00.098)       0:01:04.964 ********* 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:142
Thursday 21 July 2022  15:44:06 +0000 (0:00:00.025)       0:01:04.990 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the output when resizing with large size] *************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:148
Thursday 21 July 2022  15:44:06 +0000 (0:00:00.038)       0:01:05.028 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Remove the volume group created above] ***********************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:155
Thursday 21 July 2022  15:44:06 +0000 (0:00:00.091)       0:01:05.120 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:44:06 +0000 (0:00:00.088)       0:01:05.209 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:44:07 +0000 (0:00:00.036)       0:01:05.245 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:44:07 +0000 (0:00:00.433)       0:01:05.679 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:44:07 +0000 (0:00:00.064)       0:01:05.743 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:44:07 +0000 (0:00:00.034)       0:01:05.777 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:44:07 +0000 (0:00:00.034)       0:01:05.812 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:44:07 +0000 (0:00:00.044)       0:01:05.857 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:44:07 +0000 (0:00:00.021)       0:01:05.878 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:44:07 +0000 (0:00:00.030)       0:01:05.908 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "sdb"
            ],
            "name": "foo",
            "state": "absent",
            "type": "lvm"
        }
    ]
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:44:07 +0000 (0:00:00.036)       0:01:05.945 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:44:07 +0000 (0:00:00.035)       0:01:05.981 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:44:07 +0000 (0:00:00.027)       0:01:06.008 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:44:07 +0000 (0:00:00.025)       0:01:06.034 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:44:07 +0000 (0:00:00.027)       0:01:06.062 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:44:07 +0000 (0:00:00.030)       0:01:06.092 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:44:07 +0000 (0:00:00.060)       0:01:06.153 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:44:07 +0000 (0:00:00.023)       0:01:06.176 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "destroy device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/sdb",
            "fs_type": "lvmpv"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [
        {
            "fstype": "ext4",
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "absent"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "absent",
            "type": "lvm",
            "volumes": []
        }
    ],
    "volumes": []
}

TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91
Thursday 21 July 2022  15:44:09 +0000 (0:00:01.692)       0:01:07.868 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:44:09 +0000 (0:00:00.039)       0:01:07.907 ********* 

TASK [linux-system-roles.storage : show blivet_output] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109
Thursday 21 July 2022  15:44:09 +0000 (0:00:00.021)       0:01:07.929 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "destroy device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/sdb",
                "fs_type": "lvmpv"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [
            {
                "fstype": "ext4",
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "absent"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ],
        "volumes": []
    }
}

TASK [linux-system-roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114
Thursday 21 July 2022  15:44:09 +0000 (0:00:00.070)       0:01:08.000 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ]
    },
    "changed": false
}

TASK [linux-system-roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118
Thursday 21 July 2022  15:44:09 +0000 (0:00:00.041)       0:01:08.042 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : remove obsolete mounts] *********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134
Thursday 21 July 2022  15:44:09 +0000 (0:00:00.037)       0:01:08.079 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'ext4'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "fstype": "ext4",
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146
Thursday 21 July 2022  15:44:10 +0000 (0:00:00.385)       0:01:08.464 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : set up new/current mounts] ******************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151
Thursday 21 July 2022  15:44:10 +0000 (0:00:00.503)       0:01:08.968 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163
Thursday 21 July 2022  15:44:10 +0000 (0:00:00.042)       0:01:09.011 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171
Thursday 21 July 2022  15:44:11 +0000 (0:00:00.506)       0:01:09.517 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658418192.4109497,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176
Thursday 21 July 2022  15:44:11 +0000 (0:00:00.324)       0:01:09.841 ********* 

TASK [linux-system-roles.storage : Update facts] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198
Thursday 21 July 2022  15:44:11 +0000 (0:00:00.023)       0:01:09.865 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [Create one partition on one disk] ****************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:165
Thursday 21 July 2022  15:44:12 +0000 (0:00:00.834)       0:01:10.700 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:44:12 +0000 (0:00:00.040)       0:01:10.740 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:44:12 +0000 (0:00:00.033)       0:01:10.773 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:44:12 +0000 (0:00:00.407)       0:01:11.180 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:44:13 +0000 (0:00:00.059)       0:01:11.240 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:44:13 +0000 (0:00:00.032)       0:01:11.273 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:44:13 +0000 (0:00:00.032)       0:01:11.305 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:44:13 +0000 (0:00:00.046)       0:01:11.351 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:44:13 +0000 (0:00:00.019)       0:01:11.371 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:44:13 +0000 (0:00:00.026)       0:01:11.398 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "sdb"
            ],
            "name": "foo",
            "type": "partition",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "type": "partition"
                }
            ]
        }
    ]
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:44:13 +0000 (0:00:00.040)       0:01:11.438 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:44:13 +0000 (0:00:00.037)       0:01:11.476 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:44:13 +0000 (0:00:00.027)       0:01:11.503 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:44:13 +0000 (0:00:00.026)       0:01:11.529 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:44:13 +0000 (0:00:00.027)       0:01:11.557 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:44:13 +0000 (0:00:00.033)       0:01:11.591 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:44:13 +0000 (0:00:00.097)       0:01:11.688 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:44:13 +0000 (0:00:00.022)       0:01:11.710 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "create format",
            "device": "/dev/sdb",
            "fs_type": "disklabel"
        },
        {
            "action": "create device",
            "device": "/dev/sdb1",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/sdb1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd",
        "/dev/sdb1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext4",
            "opts": "defaults",
            "passno": 0,
            "path": "/opt/test1",
            "src": "UUID=95a42646-abe0-43b4-8c36-b1d66be20be7",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "partition",
            "volumes": [
                {
                    "_device": "/dev/sdb1",
                    "_kernel_device": "/dev/sdb1",
                    "_mount_id": "UUID=95a42646-abe0-43b4-8c36-b1d66be20be7",
                    "_raw_device": "/dev/sdb1",
                    "_raw_kernel_device": "/dev/sdb1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": 0,
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "partition",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91
Thursday 21 July 2022  15:44:15 +0000 (0:00:01.642)       0:01:13.353 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:44:15 +0000 (0:00:00.038)       0:01:13.392 ********* 

TASK [linux-system-roles.storage : show blivet_output] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109
Thursday 21 July 2022  15:44:15 +0000 (0:00:00.020)       0:01:13.412 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "create format",
                "device": "/dev/sdb",
                "fs_type": "disklabel"
            },
            {
                "action": "create device",
                "device": "/dev/sdb1",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/sdb1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd",
            "/dev/sdb1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext4",
                "opts": "defaults",
                "passno": 0,
                "path": "/opt/test1",
                "src": "UUID=95a42646-abe0-43b4-8c36-b1d66be20be7",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "partition",
                "volumes": [
                    {
                        "_device": "/dev/sdb1",
                        "_kernel_device": "/dev/sdb1",
                        "_mount_id": "UUID=95a42646-abe0-43b4-8c36-b1d66be20be7",
                        "_raw_device": "/dev/sdb1",
                        "_raw_kernel_device": "/dev/sdb1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": 0,
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "partition",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [linux-system-roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114
Thursday 21 July 2022  15:44:15 +0000 (0:00:00.066)       0:01:13.478 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "partition",
                "volumes": [
                    {
                        "_device": "/dev/sdb1",
                        "_kernel_device": "/dev/sdb1",
                        "_mount_id": "UUID=95a42646-abe0-43b4-8c36-b1d66be20be7",
                        "_raw_device": "/dev/sdb1",
                        "_raw_kernel_device": "/dev/sdb1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": 0,
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "partition",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [linux-system-roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118
Thursday 21 July 2022  15:44:15 +0000 (0:00:00.076)       0:01:13.554 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : remove obsolete mounts] *********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134
Thursday 21 July 2022  15:44:15 +0000 (0:00:00.099)       0:01:13.654 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146
Thursday 21 July 2022  15:44:15 +0000 (0:00:00.038)       0:01:13.692 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : set up new/current mounts] ******************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151
Thursday 21 July 2022  15:44:15 +0000 (0:00:00.469)       0:01:14.162 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': 'UUID=95a42646-abe0-43b4-8c36-b1d66be20be7', 'dump': 0, 'passno': 0, 'fstype': 'ext4', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "opts": "defaults",
        "passno": 0,
        "path": "/opt/test1",
        "src": "UUID=95a42646-abe0-43b4-8c36-b1d66be20be7",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "UUID=95a42646-abe0-43b4-8c36-b1d66be20be7"
}

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163
Thursday 21 July 2022  15:44:16 +0000 (0:00:00.372)       0:01:14.535 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171
Thursday 21 July 2022  15:44:16 +0000 (0:00:00.475)       0:01:15.010 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658418192.4109497,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176
Thursday 21 July 2022  15:44:17 +0000 (0:00:00.331)       0:01:15.342 ********* 

TASK [linux-system-roles.storage : Update facts] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198
Thursday 21 July 2022  15:44:17 +0000 (0:00:00.022)       0:01:15.364 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:179
Thursday 21 July 2022  15:44:18 +0000 (0:00:00.893)       0:01:16.257 ********* 
included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1
Thursday 21 July 2022  15:44:18 +0000 (0:00:00.037)       0:01:16.295 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "partition",
            "volumes": [
                {
                    "_device": "/dev/sdb1",
                    "_kernel_device": "/dev/sdb1",
                    "_mount_id": "UUID=95a42646-abe0-43b4-8c36-b1d66be20be7",
                    "_raw_device": "/dev/sdb1",
                    "_raw_kernel_device": "/dev/sdb1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": 0,
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "partition",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6
Thursday 21 July 2022  15:44:18 +0000 (0:00:00.047)       0:01:16.343 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14
Thursday 21 July 2022  15:44:18 +0000 (0:00:00.032)       0:01:16.375 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb1": {
            "fstype": "ext4",
            "label": "",
            "name": "/dev/sdb1",
            "size": "10G",
            "type": "partition",
            "uuid": "95a42646-abe0-43b4-8c36-b1d66be20be7"
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-15-42-49-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19
Thursday 21 July 2022  15:44:18 +0000 (0:00:00.327)       0:01:16.702 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.004419",
    "end": "2022-07-21 11:44:18.591311",
    "rc": 0,
    "start": "2022-07-21 11:44:18.586892"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0
UUID=95a42646-abe0-43b4-8c36-b1d66be20be7 /opt/test1 ext4 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24
Thursday 21 July 2022  15:44:18 +0000 (0:00:00.323)       0:01:17.026 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003589",
    "end": "2022-07-21 11:44:18.915358",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 11:44:18.911769"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.321)       0:01:17.348 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/sdb1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/sdb1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/sdb1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=95a42646-abe0-43b4-8c36-b1d66be20be7', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/sdb1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'partition', 'encryption_cipher': None, 'raid_spare_count': None})

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:5
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.064)       0:01:17.412 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:18
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.036)       0:01:17.448 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members)
included: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes)

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:1
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.045)       0:01:17.494 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:10
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.082)       0:01:17.577 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:19
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.023)       0:01:17.600 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:23
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.025)       0:01:17.626 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify PV count] *********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:27
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.022)       0:01:17.649 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:34
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.023)       0:01:17.673 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:38
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.021)       0:01:17.695 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:42
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.023)       0:01:17.718 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:46
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.026)       0:01:17.744 ********* 

TASK [Check MD RAID] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:56
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.023)       0:01:17.768 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-md.yml for /cache/rhel-7.qcow2

TASK [get information about RAID] **********************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:6
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.039)       0:01:17.808 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:12
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.022)       0:01:17.830 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:16
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.023)       0:01:17.853 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:20
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.021)       0:01:17.875 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:24
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.022)       0:01:17.897 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:30
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.022)       0:01:17.920 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:36
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.023)       0:01:17.943 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:44
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.023)       0:01:17.967 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:59
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.035)       0:01:18.003 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml:1
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.044)       0:01:18.047 ********* 
skipping: [/cache/rhel-7.qcow2] => (item={'_raw_device': '/dev/sdb1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/sdb1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/sdb1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=95a42646-abe0-43b4-8c36-b1d66be20be7', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/sdb1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})  => {
    "ansible_loop_var": "storage_test_lvmraid_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_lvmraid_volume": {
        "_device": "/dev/sdb1",
        "_kernel_device": "/dev/sdb1",
        "_mount_id": "UUID=95a42646-abe0-43b4-8c36-b1d66be20be7",
        "_raw_device": "/dev/sdb1",
        "_raw_kernel_device": "/dev/sdb1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [],
        "encryption": false,
        "encryption_cipher": null,
        "encryption_key": null,
        "encryption_key_size": null,
        "encryption_luks_version": null,
        "encryption_password": null,
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "ext4",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "size": 0,
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:62
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.030)       0:01:18.078 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml for /cache/rhel-7.qcow2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml:1
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.045)       0:01:18.123 ********* 
skipping: [/cache/rhel-7.qcow2] => (item={'_raw_device': '/dev/sdb1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/sdb1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/sdb1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=95a42646-abe0-43b4-8c36-b1d66be20be7', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/sdb1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})  => {
    "ansible_loop_var": "storage_test_thin_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_thin_volume": {
        "_device": "/dev/sdb1",
        "_kernel_device": "/dev/sdb1",
        "_mount_id": "UUID=95a42646-abe0-43b4-8c36-b1d66be20be7",
        "_raw_device": "/dev/sdb1",
        "_raw_kernel_device": "/dev/sdb1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [],
        "encryption": false,
        "encryption_cipher": null,
        "encryption_key": null,
        "encryption_key_size": null,
        "encryption_luks_version": null,
        "encryption_password": null,
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "ext4",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "size": 0,
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [Check member encryption] *************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:65
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.031)       0:01:18.155 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:4
Thursday 21 July 2022  15:44:19 +0000 (0:00:00.047)       0:01:18.203 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:8
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.050)       0:01:18.253 ********* 

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:15
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.021)       0:01:18.275 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:22
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.020)       0:01:18.295 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:68
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.033)       0:01:18.329 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml:1
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.046)       0:01:18.375 ********* 
skipping: [/cache/rhel-7.qcow2] => (item={'_raw_device': '/dev/sdb1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/sdb1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/sdb1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=95a42646-abe0-43b4-8c36-b1d66be20be7', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/sdb1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})  => {
    "ansible_loop_var": "storage_test_vdo_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_vdo_volume": {
        "_device": "/dev/sdb1",
        "_kernel_device": "/dev/sdb1",
        "_mount_id": "UUID=95a42646-abe0-43b4-8c36-b1d66be20be7",
        "_raw_device": "/dev/sdb1",
        "_raw_kernel_device": "/dev/sdb1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [],
        "encryption": false,
        "encryption_cipher": null,
        "encryption_key": null,
        "encryption_key_size": null,
        "encryption_luks_version": null,
        "encryption_password": null,
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "ext4",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "size": 0,
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:71
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.027)       0:01:18.403 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [verify the volumes] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml:3
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.032)       0:01:18.436 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/sdb1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/sdb1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/sdb1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=95a42646-abe0-43b4-8c36-b1d66be20be7', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/sdb1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.043)       0:01:18.479 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.051)       0:01:18.531 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache)

TASK [Get expected mount device based on device type] **************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.076)       0:01:18.607 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/sdb1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.111)       0:01:18.719 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": [
            {
                "block_available": 2402892,
                "block_size": 4096,
                "block_total": 2547269,
                "block_used": 144377,
                "device": "/dev/sdb1",
                "fstype": "ext4",
                "inode_available": 655349,
                "inode_total": 655360,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 9842245632,
                "size_total": 10433613824,
                "uuid": "95a42646-abe0-43b4-8c36-b1d66be20be7"
            }
        ],
        "storage_test_mount_expected_match_count": "1",
        "storage_test_mount_point_matches": [
            {
                "block_available": 2402892,
                "block_size": 4096,
                "block_total": 2547269,
                "block_used": 144377,
                "device": "/dev/sdb1",
                "fstype": "ext4",
                "inode_available": 655349,
                "inode_total": 655360,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 9842245632,
                "size_total": 10433613824,
                "uuid": "95a42646-abe0-43b4-8c36-b1d66be20be7"
            }
        ],
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.065)       0:01:18.784 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the current mount state by mount point] ***************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.054)       0:01:18.839 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the mount fs type] ************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.051)       0:01:18.890 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [command] *****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.049)       0:01:18.939 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.023)       0:01:18.962 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.024)       0:01:18.986 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.026)       0:01:19.012 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": null,
        "storage_test_mount_expected_match_count": null,
        "storage_test_mount_point_matches": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.034)       0:01:19.046 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "UUID=95a42646-abe0-43b4-8c36-b1d66be20be7 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.064)       0:01:19.111 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.053)       0:01:19.164 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39
Thursday 21 July 2022  15:44:20 +0000 (0:00:00.049)       0:01:19.214 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up variables] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49
Thursday 21 July 2022  15:44:21 +0000 (0:00:00.034)       0:01:19.249 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4
Thursday 21 July 2022  15:44:21 +0000 (0:00:00.035)       0:01:19.284 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10
Thursday 21 July 2022  15:44:21 +0000 (0:00:00.042)       0:01:19.327 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4
Thursday 21 July 2022  15:44:21 +0000 (0:00:00.038)       0:01:19.366 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658418254.8769498,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1658418254.8769498,
        "dev": 5,
        "device_type": 2065,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 48902,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/blockdevice",
        "mode": "0660",
        "mtime": 1658418254.8769498,
        "nlink": 1,
        "path": "/dev/sdb1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10
Thursday 21 July 2022  15:44:21 +0000 (0:00:00.324)       0:01:19.690 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18
Thursday 21 July 2022  15:44:21 +0000 (0:00:00.041)       0:01:19.732 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [(1/2) Process volume type (set initial value)] ***************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24
Thursday 21 July 2022  15:44:21 +0000 (0:00:00.041)       0:01:19.774 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "st_volume_type": "partition"
    },
    "changed": false
}

TASK [(2/2) Process volume type (get RAID value)] ******************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28
Thursday 21 July 2022  15:44:21 +0000 (0:00:00.036)       0:01:19.810 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33
Thursday 21 July 2022  15:44:21 +0000 (0:00:00.025)       0:01:19.836 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3
Thursday 21 July 2022  15:44:21 +0000 (0:00:00.041)       0:01:19.877 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10
Thursday 21 July 2022  15:44:21 +0000 (0:00:00.026)       0:01:19.904 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "rc": 0,
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.559)       0:01:20.463 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.024)       0:01:20.488 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.022)       0:01:20.511 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.049)       0:01:20.560 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.022)       0:01:20.583 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.021)       0:01:20.604 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.021)       0:01:20.626 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.021)       0:01:20.647 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.022)       0:01:20.670 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.051)       0:01:20.721 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.099)       0:01:20.821 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.038)       0:01:20.859 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.034)       0:01:20.894 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.033)       0:01:20.927 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [get information about RAID] **********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.074)       0:01:21.002 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.084)       0:01:21.087 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.037)       0:01:21.125 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.038)       0:01:21.163 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.037)       0:01:21.201 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31
Thursday 21 July 2022  15:44:22 +0000 (0:00:00.034)       0:01:21.236 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.038)       0:01:21.274 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the actual size of the volume] *************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.037)       0:01:21.311 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested size of the volume] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.027)       0:01:21.339 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.037)       0:01:21.377 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.038)       0:01:21.415 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.039)       0:01:21.454 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.038)       0:01:21.493 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.037)       0:01:21.530 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.038)       0:01:21.569 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.038)       0:01:21.607 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.040)       0:01:21.648 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.036)       0:01:21.684 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.044)       0:01:21.729 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.028)       0:01:21.757 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.029)       0:01:21.787 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check segment type] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.025)       0:01:21.813 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.024)       0:01:21.838 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested cache size] ******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.024)       0:01:21.863 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.026)       0:01:21.889 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.025)       0:01:21.914 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.024)       0:01:21.939 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.035)       0:01:21.975 ********* 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.023)       0:01:21.998 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Test setting up disk volume will remove the partition create above] ******
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:181
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.035)       0:01:22.033 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.046)       0:01:22.080 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:44:23 +0000 (0:00:00.035)       0:01:22.115 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:44:24 +0000 (0:00:00.440)       0:01:22.556 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:44:24 +0000 (0:00:00.064)       0:01:22.620 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:44:24 +0000 (0:00:00.080)       0:01:22.700 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:44:24 +0000 (0:00:00.117)       0:01:22.818 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:44:24 +0000 (0:00:00.048)       0:01:22.866 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:44:24 +0000 (0:00:00.022)       0:01:22.889 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:44:24 +0000 (0:00:00.029)       0:01:22.918 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:44:24 +0000 (0:00:00.037)       0:01:22.956 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": [
        {
            "disks": [
                "sdb"
            ],
            "fs_create_options": "-F",
            "fs_type": "ext4",
            "mount_options": "rw,noatime,defaults",
            "mount_point": "/opt/test1",
            "name": "foo",
            "type": "disk"
        }
    ]
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:44:24 +0000 (0:00:00.040)       0:01:22.996 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:44:24 +0000 (0:00:00.031)       0:01:23.027 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:44:24 +0000 (0:00:00.027)       0:01:23.055 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:44:24 +0000 (0:00:00.028)       0:01:23.083 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:44:24 +0000 (0:00:00.028)       0:01:23.112 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:44:24 +0000 (0:00:00.062)       0:01:23.175 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:44:24 +0000 (0:00:00.022)       0:01:23.197 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/sdb1",
            "fs_type": "ext4"
        },
        {
            "action": "destroy device",
            "device": "/dev/sdb1",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/sdb",
            "fs_type": "disklabel"
        },
        {
            "action": "create format",
            "device": "/dev/sdb",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [
        {
            "fstype": "ext4",
            "path": "/opt/test1",
            "src": "UUID=95a42646-abe0-43b4-8c36-b1d66be20be7",
            "state": "absent"
        },
        {
            "dump": 0,
            "fstype": "ext4",
            "opts": "rw,noatime,defaults",
            "passno": 0,
            "path": "/opt/test1",
            "src": "UUID=9fb6bd7f-4581-4dad-a181-18899aec201b",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs"
    ],
    "pools": [],
    "volumes": [
        {
            "_device": "/dev/sdb",
            "_kernel_device": "/dev/sdb",
            "_mount_id": "UUID=9fb6bd7f-4581-4dad-a181-18899aec201b",
            "_raw_device": "/dev/sdb",
            "_raw_kernel_device": "/dev/sdb",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "fs_create_options": "-F",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "ext4",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_options": "rw,noatime,defaults",
            "mount_passno": 0,
            "mount_point": "/opt/test1",
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "size": 10737418240,
            "state": "present",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91
Thursday 21 July 2022  15:44:26 +0000 (0:00:01.719)       0:01:24.916 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:44:26 +0000 (0:00:00.039)       0:01:24.956 ********* 

TASK [linux-system-roles.storage : show blivet_output] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109
Thursday 21 July 2022  15:44:26 +0000 (0:00:00.023)       0:01:24.979 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/sdb1",
                "fs_type": "ext4"
            },
            {
                "action": "destroy device",
                "device": "/dev/sdb1",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/sdb",
                "fs_type": "disklabel"
            },
            {
                "action": "create format",
                "device": "/dev/sdb",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [
            {
                "fstype": "ext4",
                "path": "/opt/test1",
                "src": "UUID=95a42646-abe0-43b4-8c36-b1d66be20be7",
                "state": "absent"
            },
            {
                "dump": 0,
                "fstype": "ext4",
                "opts": "rw,noatime,defaults",
                "passno": 0,
                "path": "/opt/test1",
                "src": "UUID=9fb6bd7f-4581-4dad-a181-18899aec201b",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs"
        ],
        "pools": [],
        "volumes": [
            {
                "_device": "/dev/sdb",
                "_kernel_device": "/dev/sdb",
                "_mount_id": "UUID=9fb6bd7f-4581-4dad-a181-18899aec201b",
                "_raw_device": "/dev/sdb",
                "_raw_kernel_device": "/dev/sdb",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "fs_create_options": "-F",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "ext4",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_options": "rw,noatime,defaults",
                "mount_passno": 0,
                "mount_point": "/opt/test1",
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "size": 10737418240,
                "state": "present",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    }
}

TASK [linux-system-roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114
Thursday 21 July 2022  15:44:26 +0000 (0:00:00.040)       0:01:25.020 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118
Thursday 21 July 2022  15:44:26 +0000 (0:00:00.036)       0:01:25.057 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": [
            {
                "_device": "/dev/sdb",
                "_kernel_device": "/dev/sdb",
                "_mount_id": "UUID=9fb6bd7f-4581-4dad-a181-18899aec201b",
                "_raw_device": "/dev/sdb",
                "_raw_kernel_device": "/dev/sdb",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "fs_create_options": "-F",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "ext4",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_options": "rw,noatime,defaults",
                "mount_passno": 0,
                "mount_point": "/opt/test1",
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "size": 10737418240,
                "state": "present",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    },
    "changed": false
}

TASK [linux-system-roles.storage : remove obsolete mounts] *********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134
Thursday 21 July 2022  15:44:26 +0000 (0:00:00.036)       0:01:25.093 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': 'UUID=95a42646-abe0-43b4-8c36-b1d66be20be7', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'ext4'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "fstype": "ext4",
        "path": "/opt/test1",
        "src": "UUID=95a42646-abe0-43b4-8c36-b1d66be20be7",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "UUID=95a42646-abe0-43b4-8c36-b1d66be20be7"
}

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146
Thursday 21 July 2022  15:44:27 +0000 (0:00:00.342)       0:01:25.435 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : set up new/current mounts] ******************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151
Thursday 21 July 2022  15:44:27 +0000 (0:00:00.454)       0:01:25.890 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': 'UUID=9fb6bd7f-4581-4dad-a181-18899aec201b', 'dump': 0, 'passno': 0, 'fstype': 'ext4', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'rw,noatime,defaults'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "opts": "rw,noatime,defaults",
        "passno": 0,
        "path": "/opt/test1",
        "src": "UUID=9fb6bd7f-4581-4dad-a181-18899aec201b",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "rw,noatime,defaults",
    "passno": "0",
    "src": "UUID=9fb6bd7f-4581-4dad-a181-18899aec201b"
}

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163
Thursday 21 July 2022  15:44:28 +0000 (0:00:00.403)       0:01:26.293 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171
Thursday 21 July 2022  15:44:28 +0000 (0:00:00.468)       0:01:26.762 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658418192.4109497,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176
Thursday 21 July 2022  15:44:28 +0000 (0:00:00.356)       0:01:27.119 ********* 

TASK [linux-system-roles.storage : Update facts] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198
Thursday 21 July 2022  15:44:28 +0000 (0:00:00.022)       0:01:27.141 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:194
Thursday 21 July 2022  15:44:29 +0000 (0:00:00.898)       0:01:28.039 ********* 
included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1
Thursday 21 July 2022  15:44:29 +0000 (0:00:00.042)       0:01:28.081 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Print out volume information] ********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6
Thursday 21 July 2022  15:44:29 +0000 (0:00:00.039)       0:01:28.121 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_volumes_list": [
        {
            "_device": "/dev/sdb",
            "_kernel_device": "/dev/sdb",
            "_mount_id": "UUID=9fb6bd7f-4581-4dad-a181-18899aec201b",
            "_raw_device": "/dev/sdb",
            "_raw_kernel_device": "/dev/sdb",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "fs_create_options": "-F",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "ext4",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_options": "rw,noatime,defaults",
            "mount_passno": 0,
            "mount_point": "/opt/test1",
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "size": 10737418240,
            "state": "present",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14
Thursday 21 July 2022  15:44:29 +0000 (0:00:00.053)       0:01:28.174 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "ext4",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": "9fb6bd7f-4581-4dad-a181-18899aec201b"
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-15-42-49-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19
Thursday 21 July 2022  15:44:30 +0000 (0:00:00.333)       0:01:28.508 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003778",
    "end": "2022-07-21 11:44:30.391253",
    "rc": 0,
    "start": "2022-07-21 11:44:30.387475"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0
UUID=9fb6bd7f-4581-4dad-a181-18899aec201b /opt/test1 ext4 rw,noatime,defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24
Thursday 21 July 2022  15:44:30 +0000 (0:00:00.314)       0:01:28.822 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003469",
    "end": "2022-07-21 11:44:30.696989",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 11:44:30.693520"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33
Thursday 21 July 2022  15:44:30 +0000 (0:00:00.305)       0:01:29.127 ********* 

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43
Thursday 21 July 2022  15:44:30 +0000 (0:00:00.022)       0:01:29.150 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/sdb', 'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'ext4', 'mount_options': 'rw,noatime,defaults', '_device': '/dev/sdb', 'size': 10737418240, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/sdb', 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'type': 'disk', 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=9fb6bd7f-4581-4dad-a181-18899aec201b', 'raid_spare_count': None, 'name': 'foo', '_raw_kernel_device': '/dev/sdb', 'cache_mode': None, 'cache_devices': [], 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': None, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'thin_pool_size': None, 'fs_create_options': '-F'})

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2
Thursday 21 July 2022  15:44:30 +0000 (0:00:00.057)       0:01:29.207 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10
Thursday 21 July 2022  15:44:31 +0000 (0:00:00.050)       0:01:29.257 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache)

TASK [Get expected mount device based on device type] **************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6
Thursday 21 July 2022  15:44:31 +0000 (0:00:00.074)       0:01:29.332 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/sdb"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14
Thursday 21 July 2022  15:44:31 +0000 (0:00:00.039)       0:01:29.372 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": [
            {
                "block_available": 2403135,
                "block_size": 4096,
                "block_total": 2547525,
                "block_used": 144390,
                "device": "/dev/sdb",
                "fstype": "ext4",
                "inode_available": 655349,
                "inode_total": 655360,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,noatime,data=ordered",
                "size_available": 9843240960,
                "size_total": 10434662400,
                "uuid": "9fb6bd7f-4581-4dad-a181-18899aec201b"
            }
        ],
        "storage_test_mount_expected_match_count": "1",
        "storage_test_mount_point_matches": [
            {
                "block_available": 2403135,
                "block_size": 4096,
                "block_total": 2547525,
                "block_used": 144390,
                "device": "/dev/sdb",
                "fstype": "ext4",
                "inode_available": 655349,
                "inode_total": 655360,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,noatime,data=ordered",
                "size_available": 9843240960,
                "size_total": 10434662400,
                "uuid": "9fb6bd7f-4581-4dad-a181-18899aec201b"
            }
        ],
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28
Thursday 21 July 2022  15:44:31 +0000 (0:00:00.095)       0:01:29.468 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the current mount state by mount point] ***************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37
Thursday 21 July 2022  15:44:31 +0000 (0:00:00.054)       0:01:29.522 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the mount fs type] ************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45
Thursday 21 July 2022  15:44:31 +0000 (0:00:00.095)       0:01:29.617 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [command] *****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54
Thursday 21 July 2022  15:44:31 +0000 (0:00:00.090)       0:01:29.708 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58
Thursday 21 July 2022  15:44:31 +0000 (0:00:00.063)       0:01:29.772 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63
Thursday 21 July 2022  15:44:31 +0000 (0:00:00.023)       0:01:29.795 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75
Thursday 21 July 2022  15:44:31 +0000 (0:00:00.022)       0:01:29.817 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": null,
        "storage_test_mount_expected_match_count": null,
        "storage_test_mount_point_matches": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2
Thursday 21 July 2022  15:44:31 +0000 (0:00:00.037)       0:01:29.855 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "UUID=9fb6bd7f-4581-4dad-a181-18899aec201b "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 rw,noatime,defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25
Thursday 21 July 2022  15:44:31 +0000 (0:00:00.061)       0:01:29.917 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32
Thursday 21 July 2022  15:44:31 +0000 (0:00:00.050)       0:01:29.967 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39
Thursday 21 July 2022  15:44:31 +0000 (0:00:00.049)       0:01:30.017 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49
Thursday 21 July 2022  15:44:31 +0000 (0:00:00.060)       0:01:30.078 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4
Thursday 21 July 2022  15:44:31 +0000 (0:00:00.035)       0:01:30.113 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10
Thursday 21 July 2022  15:44:31 +0000 (0:00:00.037)       0:01:30.151 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4
Thursday 21 July 2022  15:44:31 +0000 (0:00:00.038)       0:01:30.189 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658418266.4539497,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1658418266.4539497,
        "dev": 5,
        "device_type": 2064,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 2032,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/blockdevice",
        "mode": "0660",
        "mtime": 1658418266.4539497,
        "nlink": 1,
        "path": "/dev/sdb",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10
Thursday 21 July 2022  15:44:32 +0000 (0:00:00.339)       0:01:30.529 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18
Thursday 21 July 2022  15:44:32 +0000 (0:00:00.040)       0:01:30.569 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [(1/2) Process volume type (set initial value)] ***************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24
Thursday 21 July 2022  15:44:32 +0000 (0:00:00.041)       0:01:30.610 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "st_volume_type": "disk"
    },
    "changed": false
}

TASK [(2/2) Process volume type (get RAID value)] ******************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28
Thursday 21 July 2022  15:44:32 +0000 (0:00:00.039)       0:01:30.650 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33
Thursday 21 July 2022  15:44:32 +0000 (0:00:00.024)       0:01:30.674 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3
Thursday 21 July 2022  15:44:32 +0000 (0:00:00.040)       0:01:30.715 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10
Thursday 21 July 2022  15:44:32 +0000 (0:00:00.024)       0:01:30.739 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "rc": 0,
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.573)       0:01:31.313 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.025)       0:01:31.339 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.025)       0:01:31.365 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.052)       0:01:31.417 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.023)       0:01:31.441 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.023)       0:01:31.465 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.022)       0:01:31.487 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.021)       0:01:31.508 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.022)       0:01:31.531 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.051)       0:01:31.583 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.050)       0:01:31.633 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.034)       0:01:31.667 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.036)       0:01:31.704 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.040)       0:01:31.745 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [get information about RAID] **********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.032)       0:01:31.777 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.035)       0:01:31.812 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.033)       0:01:31.846 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.038)       0:01:31.884 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.037)       0:01:31.921 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.034)       0:01:31.956 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.035)       0:01:31.992 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the actual size of the volume] *************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.042)       0:01:32.034 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested size of the volume] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.024)       0:01:32.059 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.036)       0:01:32.096 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20
Thursday 21 July 2022  15:44:33 +0000 (0:00:00.034)       0:01:32.131 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.124)       0:01:32.255 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.037)       0:01:32.293 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.038)       0:01:32.331 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.037)       0:01:32.369 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.044)       0:01:32.413 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.036)       0:01:32.449 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.036)       0:01:32.486 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.036)       0:01:32.522 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.024)       0:01:32.546 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.024)       0:01:32.571 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check segment type] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.025)       0:01:32.597 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.025)       0:01:32.622 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested cache size] ******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.026)       0:01:32.648 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.023)       0:01:32.672 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.022)       0:01:32.695 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.022)       0:01:32.718 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.034)       0:01:32.752 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Remove the disk volume created above] ************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:198
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.031)       0:01:32.784 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.052)       0:01:32.837 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:44:34 +0000 (0:00:00.034)       0:01:32.871 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:44:35 +0000 (0:00:00.409)       0:01:33.281 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:44:35 +0000 (0:00:00.059)       0:01:33.341 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:44:35 +0000 (0:00:00.032)       0:01:33.374 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:44:35 +0000 (0:00:00.033)       0:01:33.407 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:44:35 +0000 (0:00:00.046)       0:01:33.453 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:44:35 +0000 (0:00:00.020)       0:01:33.474 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:44:35 +0000 (0:00:00.031)       0:01:33.506 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:44:35 +0000 (0:00:00.036)       0:01:33.542 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": [
        {
            "disks": [
                "sdb"
            ],
            "name": "foo",
            "state": "absent",
            "type": "disk"
        }
    ]
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:44:35 +0000 (0:00:00.037)       0:01:33.579 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:44:35 +0000 (0:00:00.029)       0:01:33.609 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:44:35 +0000 (0:00:00.032)       0:01:33.641 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:44:35 +0000 (0:00:00.028)       0:01:33.669 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:44:35 +0000 (0:00:00.028)       0:01:33.697 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:44:35 +0000 (0:00:00.058)       0:01:33.756 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:44:35 +0000 (0:00:00.020)       0:01:33.777 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/sdb",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [
        {
            "fstype": "ext4",
            "path": "/opt/test1",
            "src": "UUID=9fb6bd7f-4581-4dad-a181-18899aec201b",
            "state": "absent"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [],
    "volumes": [
        {
            "_device": "/dev/sdb",
            "_mount_id": "UUID=9fb6bd7f-4581-4dad-a181-18899aec201b",
            "_raw_device": "/dev/sdb",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "fs_create_options": "",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "ext4",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_options": "defaults",
            "mount_passno": 0,
            "mount_point": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "size": 10737418240,
            "state": "absent",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91
Thursday 21 July 2022  15:44:36 +0000 (0:00:01.119)       0:01:34.896 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:44:36 +0000 (0:00:00.035)       0:01:34.931 ********* 

TASK [linux-system-roles.storage : show blivet_output] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109
Thursday 21 July 2022  15:44:36 +0000 (0:00:00.021)       0:01:34.953 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/sdb",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [
            {
                "fstype": "ext4",
                "path": "/opt/test1",
                "src": "UUID=9fb6bd7f-4581-4dad-a181-18899aec201b",
                "state": "absent"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [],
        "volumes": [
            {
                "_device": "/dev/sdb",
                "_mount_id": "UUID=9fb6bd7f-4581-4dad-a181-18899aec201b",
                "_raw_device": "/dev/sdb",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "fs_create_options": "",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "ext4",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_options": "defaults",
                "mount_passno": 0,
                "mount_point": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "size": 10737418240,
                "state": "absent",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    }
}

TASK [linux-system-roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114
Thursday 21 July 2022  15:44:36 +0000 (0:00:00.078)       0:01:35.032 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118
Thursday 21 July 2022  15:44:36 +0000 (0:00:00.096)       0:01:35.129 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": [
            {
                "_device": "/dev/sdb",
                "_mount_id": "UUID=9fb6bd7f-4581-4dad-a181-18899aec201b",
                "_raw_device": "/dev/sdb",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "fs_create_options": "",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "ext4",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_options": "defaults",
                "mount_passno": 0,
                "mount_point": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "size": 10737418240,
                "state": "absent",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    },
    "changed": false
}

TASK [linux-system-roles.storage : remove obsolete mounts] *********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134
Thursday 21 July 2022  15:44:36 +0000 (0:00:00.039)       0:01:35.168 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': 'UUID=9fb6bd7f-4581-4dad-a181-18899aec201b', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'ext4'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "fstype": "ext4",
        "path": "/opt/test1",
        "src": "UUID=9fb6bd7f-4581-4dad-a181-18899aec201b",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "UUID=9fb6bd7f-4581-4dad-a181-18899aec201b"
}

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146
Thursday 21 July 2022  15:44:37 +0000 (0:00:00.356)       0:01:35.525 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : set up new/current mounts] ******************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151
Thursday 21 July 2022  15:44:37 +0000 (0:00:00.471)       0:01:35.996 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163
Thursday 21 July 2022  15:44:37 +0000 (0:00:00.040)       0:01:36.037 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171
Thursday 21 July 2022  15:44:38 +0000 (0:00:00.447)       0:01:36.484 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658418192.4109497,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176
Thursday 21 July 2022  15:44:38 +0000 (0:00:00.328)       0:01:36.813 ********* 

TASK [linux-system-roles.storage : Update facts] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198
Thursday 21 July 2022  15:44:38 +0000 (0:00:00.023)       0:01:36.836 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [Try to mount swap filesystem to "/opt/test1"] ****************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:210
Thursday 21 July 2022  15:44:39 +0000 (0:00:00.843)       0:01:37.679 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:44:39 +0000 (0:00:00.037)       0:01:37.716 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:44:39 +0000 (0:00:00.032)       0:01:37.749 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:44:39 +0000 (0:00:00.418)       0:01:38.168 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:44:39 +0000 (0:00:00.063)       0:01:38.231 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:44:40 +0000 (0:00:00.037)       0:01:38.269 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:44:40 +0000 (0:00:00.034)       0:01:38.304 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:44:40 +0000 (0:00:00.046)       0:01:38.350 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:44:40 +0000 (0:00:00.022)       0:01:38.373 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:44:40 +0000 (0:00:00.026)       0:01:38.400 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:44:40 +0000 (0:00:00.071)       0:01:38.471 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": [
        {
            "disks": [
                "sdb"
            ],
            "fs_type": "swap",
            "mount_point": "/opt/test1",
            "name": "test1",
            "type": "disk"
        }
    ]
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:44:40 +0000 (0:00:00.037)       0:01:38.508 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:44:40 +0000 (0:00:00.028)       0:01:38.537 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:44:40 +0000 (0:00:00.026)       0:01:38.564 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:44:40 +0000 (0:00:00.025)       0:01:38.589 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:44:40 +0000 (0:00:00.026)       0:01:38.615 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:44:40 +0000 (0:00:00.053)       0:01:38.668 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:44:40 +0000 (0:00:00.021)       0:01:38.690 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

volume 'test1' has a mount point but no mountable file system

TASK [linux-system-roles.storage : failed message] *****************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99
Thursday 21 July 2022  15:44:41 +0000 (0:00:00.908)       0:01:39.599 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "changed": false
}

MSG:

{'crypts': [], 'mounts': [], 'leaves': [], 'changed': False, 'actions': [], 'failed': True, 'volumes': [], 'invocation': {'module_args': {'packages_only': False, 'disklabel_type': None, 'diskvolume_mkfs_option_map': {'ext4': '-F', 'ext3': '-F', 'ext2': '-F'}, 'safe_mode': False, 'pools': [], 'volumes': [{'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'swap', 'mount_options': 'defaults', 'size': 10737418240, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'type': 'disk', 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, 'raid_spare_count': None, 'name': 'test1', 'cache_mode': None, 'cache_devices': [], 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': None, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'thin_pool_size': None, 'fs_create_options': ''}], 'pool_defaults': {'encryption_password': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_cipher': None, 'disks': [], 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_device_count': None, 'state': 'present', 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'raid_level': None, 'raid_spare_count': None}, 'volume_defaults': {'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'xfs', 'mount_options': 'defaults', 'size': 0, 'mount_point': '', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'fs_overwrite_existing': True, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'mount_passno': 0, 'raid_spare_count': None, 'cache_mode': None, 'deduplication': None, 'cached': False, 'type': 'lvm', 'disks': [], 'thin_pool_size': None, 'thin': None, 'mount_check': 0, 'cache_size': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, 'use_partitions': None}}, 'pools': [], 'packages': [], 'msg': "volume 'test1' has a mount point but no mountable file system", '_ansible_no_log': False}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:44:41 +0000 (0:00:00.041)       0:01:39.640 ********* 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:226
Thursday 21 July 2022  15:44:41 +0000 (0:00:00.022)       0:01:39.662 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the output when mount swap filesystem to "/opt/test1"] ************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:232
Thursday 21 July 2022  15:44:41 +0000 (0:00:00.034)       0:01:39.697 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed
META: ran handlers
META: ran handlers

PLAY RECAP *********************************************************************
/cache/rhel-7.qcow2        : ok=483  changed=19   unreachable=0    failed=3    skipped=367  rescued=3    ignored=0   

Thursday 21 July 2022  15:44:41 +0000 (0:00:00.059)       0:01:39.757 ********* 
=============================================================================== 
linux-system-roles.storage : make sure blivet is available -------------- 9.12s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.72s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.69s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.69s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.64s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.62s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.60s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
Ensure cryptsetup is present -------------------------------------------- 1.56s
/tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 -------------------
set up internal repositories -------------------------------------------- 1.44s
/cache/rhel-7_setup.yml:5 -----------------------------------------------------
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.36s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.33s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
Collect info about the volumes. ----------------------------------------- 1.32s
/tmp/tmpaxjje44y/tests/verify-role-results.yml:14 -----------------------------
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.31s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
linux-system-roles.storage : get service facts -------------------------- 1.19s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.12s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
Gathering Facts --------------------------------------------------------- 1.04s
/tmp/tmpaxjje44y/tests/tests_misc_scsi_generated.yml:3 ------------------------
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 0.91s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
linux-system-roles.storage : Update facts ------------------------------- 0.90s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 
linux-system-roles.storage : Update facts ------------------------------- 0.89s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 
linux-system-roles.storage : Update facts ------------------------------- 0.86s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 
ansible-playbook [core 2.12.6]
  config file = /etc/ansible/ansible.cfg
  configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
  ansible python module location = /usr/lib/python3.9/site-packages/ansible
  ansible collection location = /tmp/tmp5bkr4li_
  executable location = /usr/bin/ansible-playbook
  python version = 3.9.13 (main, May 18 2022, 00:00:00) [GCC 11.3.1 20220421 (Red Hat 11.3.1-2)]
  jinja version = 2.11.3
  libyaml = True
Using /etc/ansible/ansible.cfg as config file
Skipping callback 'debug', as we already have a stdout callback.
Skipping callback 'default', as we already have a stdout callback.
Skipping callback 'minimal', as we already have a stdout callback.
Skipping callback 'oneline', as we already have a stdout callback.

PLAYBOOK: rhel-7_setup.yml *****************************************************
1 plays in /cache/rhel-7_setup.yml

PLAY [Setup repos] *************************************************************
META: ran handlers

TASK [set up internal repositories] ********************************************
task path: /cache/rhel-7_setup.yml:5
Thursday 21 July 2022  18:52:28 +0000 (0:00:00.018)       0:00:00.018 ********* 
changed: [/cache/rhel-7.qcow2] => (item=None) => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
changed: [/cache/rhel-7.qcow2] => (item=None) => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
changed: [/cache/rhel-7.qcow2] => (item=None) => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
changed: [/cache/rhel-7.qcow2] => (item=None) => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
changed: [/cache/rhel-7.qcow2] => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
META: ran handlers
META: ran handlers

PLAY RECAP *********************************************************************
/cache/rhel-7.qcow2        : ok=1    changed=1    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   

Thursday 21 July 2022  18:52:30 +0000 (0:00:01.422)       0:00:01.440 ********* 
=============================================================================== 
set up internal repositories -------------------------------------------- 1.42s
/cache/rhel-7_setup.yml:5 -----------------------------------------------------

PLAYBOOK: tests_misc_scsi_generated.yml ****************************************
2 plays in /tmp/tmptomayb7j/tests/storage/tests_misc_scsi_generated.yml

PLAY [all] *********************************************************************

TASK [Gathering Facts] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc_scsi_generated.yml:3
Thursday 21 July 2022  18:52:30 +0000 (0:00:00.022)       0:00:01.462 ********* 
ok: [/cache/rhel-7.qcow2]
META: ran handlers

TASK [set disk interface for test] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc_scsi_generated.yml:7
Thursday 21 July 2022  18:52:31 +0000 (0:00:01.030)       0:00:02.492 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_use_interface": "scsi"
    },
    "changed": false
}
META: ran handlers
META: ran handlers

PLAY [all] *********************************************************************

TASK [Gathering Facts] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:2
Thursday 21 July 2022  18:52:31 +0000 (0:00:00.048)       0:00:02.540 ********* 
ok: [/cache/rhel-7.qcow2]
META: ran handlers

TASK [include_role : fedora.linux_system_roles.storage] ************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:15
Thursday 21 July 2022  18:52:31 +0000 (0:00:00.756)       0:00:03.296 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:52:31 +0000 (0:00:00.036)       0:00:03.332 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:52:32 +0000 (0:00:00.028)       0:00:03.361 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:52:32 +0000 (0:00:00.413)       0:00:03.774 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:52:32 +0000 (0:00:00.059)       0:00:03.834 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:52:32 +0000 (0:00:00.033)       0:00:03.868 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:52:32 +0000 (0:00:00.030)       0:00:03.898 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:52:32 +0000 (0:00:00.056)       0:00:03.955 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:52:32 +0000 (0:00:00.019)       0:00:03.974 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "changed": true,
    "changes": {
        "installed": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "rc": 0,
    "results": [
        "Loaded plugins: search-disabled-repos\nResolving Dependencies\n--> Running transaction check\n---> Package libblockdev-crypto.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libblockdev-utils(x86-64) = 2.18-5.el7 for package: libblockdev-crypto-2.18-5.el7.x86_64\n--> Processing Dependency: libvolume_key.so.1()(64bit) for package: libblockdev-crypto-2.18-5.el7.x86_64\n--> Processing Dependency: libbd_utils.so.2()(64bit) for package: libblockdev-crypto-2.18-5.el7.x86_64\n---> Package libblockdev-dm.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libdmraid.so.1(Base)(64bit) for package: libblockdev-dm-2.18-5.el7.x86_64\n--> Processing Dependency: dmraid for package: libblockdev-dm-2.18-5.el7.x86_64\n--> Processing Dependency: libdmraid.so.1()(64bit) for package: libblockdev-dm-2.18-5.el7.x86_64\n---> Package libblockdev-lvm.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: lvm2 for package: libblockdev-lvm-2.18-5.el7.x86_64\n--> Processing Dependency: device-mapper-persistent-data for package: libblockdev-lvm-2.18-5.el7.x86_64\n---> Package libblockdev-mdraid.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: mdadm for package: libblockdev-mdraid-2.18-5.el7.x86_64\n--> Processing Dependency: libbytesize.so.1()(64bit) for package: libblockdev-mdraid-2.18-5.el7.x86_64\n---> Package libblockdev-swap.x86_64 0:2.18-5.el7 will be installed\n---> Package python-enum34.noarch 0:1.0.4-1.el7 will be installed\n---> Package python2-blivet3.noarch 1:3.1.3-3.el7 will be installed\n--> Processing Dependency: blivet3-data = 1:3.1.3-3.el7 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-bytesize >= 0.3 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-blockdev >= 2.17 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: pyparted >= 3.9 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-hawkey for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: lsof for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Running transaction check\n---> Package blivet3-data.noarch 1:3.1.3-3.el7 will be installed\n---> Package device-mapper-persistent-data.x86_64 0:0.8.5-3.el7_9.2 will be installed\n--> Processing Dependency: libaio.so.1(LIBAIO_0.4)(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n--> Processing Dependency: libaio.so.1(LIBAIO_0.1)(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n--> Processing Dependency: libaio.so.1()(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n---> Package dmraid.x86_64 0:1.0.0.rc16-28.el7 will be installed\n--> Processing Dependency: libdevmapper-event.so.1.02(Base)(64bit) for package: dmraid-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: dmraid-events for package: dmraid-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: libdevmapper-event.so.1.02()(64bit) for package: dmraid-1.0.0.rc16-28.el7.x86_64\n---> Package libblockdev-utils.x86_64 0:2.18-5.el7 will be installed\n---> Package libbytesize.x86_64 0:1.2-1.el7 will be installed\n--> Processing Dependency: libmpfr.so.4()(64bit) for package: libbytesize-1.2-1.el7.x86_64\n---> Package lsof.x86_64 0:4.87-6.el7 will be installed\n---> Package lvm2.x86_64 7:2.02.187-6.el7_9.5 will be installed\n--> Processing Dependency: lvm2-libs = 7:2.02.187-6.el7_9.5 for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n--> Processing Dependency: liblvm2app.so.2.2(Base)(64bit) for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n--> Processing Dependency: liblvm2app.so.2.2()(64bit) for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n---> Package mdadm.x86_64 0:4.1-9.el7_9 will be installed\n--> Processing Dependency: libreport-filesystem for package: mdadm-4.1-9.el7_9.x86_64\n---> Package pyparted.x86_64 1:3.9-15.el7 will be installed\n---> Package python2-blockdev.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libblockdev(x86-64) = 2.18-5.el7 for package: python2-blockdev-2.18-5.el7.x86_64\n---> Package python2-bytesize.x86_64 0:1.2-1.el7 will be installed\n---> Package python2-hawkey.x86_64 0:0.22.5-2.el7_9 will be installed\n--> Processing Dependency: libdnf(x86-64) = 0.22.5-2.el7_9 for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: python2-libdnf = 0.22.5-2.el7_9 for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolv.so.0(SOLV_1.0)(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolvext.so.0(SOLV_1.0)(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libdnf.so.2()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libjson-glib-1.0.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libmodulemd.so.1()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: librepo.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: librhsm.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolv.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolvext.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n---> Package volume_key-libs.x86_64 0:0.3.9-9.el7 will be installed\n--> Running transaction check\n---> Package device-mapper-event-libs.x86_64 7:1.02.170-6.el7_9.5 will be installed\n---> Package dmraid-events.x86_64 0:1.0.0.rc16-28.el7 will be installed\n--> Processing Dependency: sgpio for package: dmraid-events-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: device-mapper-event for package: dmraid-events-1.0.0.rc16-28.el7.x86_64\n---> Package json-glib.x86_64 0:1.4.2-2.el7 will be installed\n---> Package libaio.x86_64 0:0.3.109-13.el7 will be installed\n---> Package libblockdev.x86_64 0:2.18-5.el7 will be installed\n---> Package libdnf.x86_64 0:0.22.5-2.el7_9 will be installed\n---> Package libmodulemd.x86_64 0:1.6.3-1.el7 will be installed\n---> Package librepo.x86_64 0:1.8.1-8.el7_9 will be installed\n---> Package libreport-filesystem.x86_64 0:2.1.11-53.el7 will be installed\n---> Package librhsm.x86_64 0:0.0.3-3.el7_9 will be installed\n---> Package libsolv.x86_64 0:0.6.34-4.el7 will be installed\n---> Package lvm2-libs.x86_64 7:2.02.187-6.el7_9.5 will be installed\n---> Package mpfr.x86_64 0:3.1.1-4.el7 will be installed\n---> Package python2-libdnf.x86_64 0:0.22.5-2.el7_9 will be installed\n--> Running transaction check\n---> Package device-mapper-event.x86_64 7:1.02.170-6.el7_9.5 will be installed\n---> Package sgpio.x86_64 0:1.2.0.10-13.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package            Arch   Version              Repository                 Size\n================================================================================\nInstalling:\n libblockdev-crypto x86_64 2.18-5.el7           rhel                       60 k\n libblockdev-dm     x86_64 2.18-5.el7           rhel-optional              54 k\n libblockdev-lvm    x86_64 2.18-5.el7           rhel                       62 k\n libblockdev-mdraid x86_64 2.18-5.el7           rhel                       57 k\n libblockdev-swap   x86_64 2.18-5.el7           rhel                       52 k\n python-enum34      noarch 1.0.4-1.el7          rhel                       52 k\n python2-blivet3    noarch 1:3.1.3-3.el7        rhel                      851 k\nInstalling for dependencies:\n blivet3-data       noarch 1:3.1.3-3.el7        rhel                       77 k\n device-mapper-event\n                    x86_64 7:1.02.170-6.el7_9.5 rhel                      192 k\n device-mapper-event-libs\n                    x86_64 7:1.02.170-6.el7_9.5 rhel                      192 k\n device-mapper-persistent-data\n                    x86_64 0.8.5-3.el7_9.2      rhel                      423 k\n dmraid             x86_64 1.0.0.rc16-28.el7    rhel                      151 k\n dmraid-events      x86_64 1.0.0.rc16-28.el7    rhel                       21 k\n json-glib          x86_64 1.4.2-2.el7          rhel                      134 k\n libaio             x86_64 0.3.109-13.el7       rhel                       24 k\n libblockdev        x86_64 2.18-5.el7           rhel                      119 k\n libblockdev-utils  x86_64 2.18-5.el7           rhel                       59 k\n libbytesize        x86_64 1.2-1.el7            rhel                       52 k\n libdnf             x86_64 0.22.5-2.el7_9       rhel-7-server-extras-rpms 536 k\n libmodulemd        x86_64 1.6.3-1.el7          rhel-7-server-extras-rpms 153 k\n librepo            x86_64 1.8.1-8.el7_9        rhel                       82 k\n libreport-filesystem\n                    x86_64 2.1.11-53.el7        rhel                       41 k\n librhsm            x86_64 0.0.3-3.el7_9        rhel-7-server-extras-rpms  28 k\n libsolv            x86_64 0.6.34-4.el7         rhel                      329 k\n lsof               x86_64 4.87-6.el7           rhel                      331 k\n lvm2               x86_64 7:2.02.187-6.el7_9.5 rhel                      1.3 M\n lvm2-libs          x86_64 7:2.02.187-6.el7_9.5 rhel                      1.1 M\n mdadm              x86_64 4.1-9.el7_9          rhel                      440 k\n mpfr               x86_64 3.1.1-4.el7          rhel                      203 k\n pyparted           x86_64 1:3.9-15.el7         rhel                      195 k\n python2-blockdev   x86_64 2.18-5.el7           rhel                       61 k\n python2-bytesize   x86_64 1.2-1.el7            rhel                       22 k\n python2-hawkey     x86_64 0.22.5-2.el7_9       rhel-7-server-extras-rpms  71 k\n python2-libdnf     x86_64 0.22.5-2.el7_9       rhel-7-server-extras-rpms 611 k\n sgpio              x86_64 1.2.0.10-13.el7      rhel                       14 k\n volume_key-libs    x86_64 0.3.9-9.el7          rhel                      141 k\n\nTransaction Summary\n================================================================================\nInstall  7 Packages (+29 Dependent packages)\n\nTotal download size: 8.2 M\nInstalled size: 24 M\nDownloading packages:\n--------------------------------------------------------------------------------\nTotal                                               19 MB/s | 8.2 MB  00:00     \nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n  Installing : libblockdev-utils-2.18-5.el7.x86_64                         1/36 \n  Installing : 7:device-mapper-event-libs-1.02.170-6.el7_9.5.x86_64        2/36 \n  Installing : json-glib-1.4.2-2.el7.x86_64                                3/36 \n  Installing : librhsm-0.0.3-3.el7_9.x86_64                                4/36 \n  Installing : libsolv-0.6.34-4.el7.x86_64                                 5/36 \n  Installing : libaio-0.3.109-13.el7.x86_64                                6/36 \n  Installing : librepo-1.8.1-8.el7_9.x86_64                                7/36 \n  Installing : libmodulemd-1.6.3-1.el7.x86_64                              8/36 \n  Installing : libdnf-0.22.5-2.el7_9.x86_64                                9/36 \n  Installing : device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64       10/36 \n  Installing : 7:device-mapper-event-1.02.170-6.el7_9.5.x86_64            11/36 \n  Installing : 7:lvm2-libs-2.02.187-6.el7_9.5.x86_64                      12/36 \n  Installing : 7:lvm2-2.02.187-6.el7_9.5.x86_64                           13/36 \n  Installing : python2-libdnf-0.22.5-2.el7_9.x86_64                       14/36 \n  Installing : python2-hawkey-0.22.5-2.el7_9.x86_64                       15/36 \n  Installing : libblockdev-2.18-5.el7.x86_64                              16/36 \n  Installing : python2-blockdev-2.18-5.el7.x86_64                         17/36 \n  Installing : 1:pyparted-3.9-15.el7.x86_64                               18/36 \n  Installing : sgpio-1.2.0.10-13.el7.x86_64                               19/36 \n  Installing : dmraid-1.0.0.rc16-28.el7.x86_64                            20/36 \n  Installing : dmraid-events-1.0.0.rc16-28.el7.x86_64                     21/36 \n  Installing : volume_key-libs-0.3.9-9.el7.x86_64                         22/36 \n  Installing : mpfr-3.1.1-4.el7.x86_64                                    23/36 \n  Installing : libbytesize-1.2-1.el7.x86_64                               24/36 \n  Installing : python2-bytesize-1.2-1.el7.x86_64                          25/36 \n  Installing : libreport-filesystem-2.1.11-53.el7.x86_64                  26/36 \n  Installing : mdadm-4.1-9.el7_9.x86_64                                   27/36 \n  Installing : 1:blivet3-data-3.1.3-3.el7.noarch                          28/36 \n  Installing : lsof-4.87-6.el7.x86_64                                     29/36 \n  Installing : 1:python2-blivet3-3.1.3-3.el7.noarch                       30/36 \n  Installing : libblockdev-mdraid-2.18-5.el7.x86_64                       31/36 \n  Installing : libblockdev-crypto-2.18-5.el7.x86_64                       32/36 \n  Installing : libblockdev-dm-2.18-5.el7.x86_64                           33/36 \n  Installing : libblockdev-lvm-2.18-5.el7.x86_64                          34/36 \n  Installing : libblockdev-swap-2.18-5.el7.x86_64                         35/36 \n  Installing : python-enum34-1.0.4-1.el7.noarch                           36/36 \n  Verifying  : 7:device-mapper-event-1.02.170-6.el7_9.5.x86_64             1/36 \n  Verifying  : libblockdev-swap-2.18-5.el7.x86_64                          2/36 \n  Verifying  : librhsm-0.0.3-3.el7_9.x86_64                                3/36 \n  Verifying  : libblockdev-lvm-2.18-5.el7.x86_64                           4/36 \n  Verifying  : lsof-4.87-6.el7.x86_64                                      5/36 \n  Verifying  : libblockdev-mdraid-2.18-5.el7.x86_64                        6/36 \n  Verifying  : libdnf-0.22.5-2.el7_9.x86_64                                7/36 \n  Verifying  : python-enum34-1.0.4-1.el7.noarch                            8/36 \n  Verifying  : 1:blivet3-data-3.1.3-3.el7.noarch                           9/36 \n  Verifying  : dmraid-events-1.0.0.rc16-28.el7.x86_64                     10/36 \n  Verifying  : python2-blockdev-2.18-5.el7.x86_64                         11/36 \n  Verifying  : libmodulemd-1.6.3-1.el7.x86_64                             12/36 \n  Verifying  : librepo-1.8.1-8.el7_9.x86_64                               13/36 \n  Verifying  : libblockdev-dm-2.18-5.el7.x86_64                           14/36 \n  Verifying  : json-glib-1.4.2-2.el7.x86_64                               15/36 \n  Verifying  : libaio-0.3.109-13.el7.x86_64                               16/36 \n  Verifying  : 7:lvm2-libs-2.02.187-6.el7_9.5.x86_64                      17/36 \n  Verifying  : python2-hawkey-0.22.5-2.el7_9.x86_64                       18/36 \n  Verifying  : python2-bytesize-1.2-1.el7.x86_64                          19/36 \n  Verifying  : libblockdev-2.18-5.el7.x86_64                              20/36 \n  Verifying  : libreport-filesystem-2.1.11-53.el7.x86_64                  21/36 \n  Verifying  : libbytesize-1.2-1.el7.x86_64                               22/36 \n  Verifying  : 7:device-mapper-event-libs-1.02.170-6.el7_9.5.x86_64       23/36 \n  Verifying  : python2-libdnf-0.22.5-2.el7_9.x86_64                       24/36 \n  Verifying  : 7:lvm2-2.02.187-6.el7_9.5.x86_64                           25/36 \n  Verifying  : libblockdev-utils-2.18-5.el7.x86_64                        26/36 \n  Verifying  : mpfr-3.1.1-4.el7.x86_64                                    27/36 \n  Verifying  : volume_key-libs-0.3.9-9.el7.x86_64                         28/36 \n  Verifying  : libsolv-0.6.34-4.el7.x86_64                                29/36 \n  Verifying  : device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64       30/36 \n  Verifying  : 1:python2-blivet3-3.1.3-3.el7.noarch                       31/36 \n  Verifying  : dmraid-1.0.0.rc16-28.el7.x86_64                            32/36 \n  Verifying  : mdadm-4.1-9.el7_9.x86_64                                   33/36 \n  Verifying  : sgpio-1.2.0.10-13.el7.x86_64                               34/36 \n  Verifying  : libblockdev-crypto-2.18-5.el7.x86_64                       35/36 \n  Verifying  : 1:pyparted-3.9-15.el7.x86_64                               36/36 \n\nInstalled:\n  libblockdev-crypto.x86_64 0:2.18-5.el7 libblockdev-dm.x86_64 0:2.18-5.el7    \n  libblockdev-lvm.x86_64 0:2.18-5.el7    libblockdev-mdraid.x86_64 0:2.18-5.el7\n  libblockdev-swap.x86_64 0:2.18-5.el7   python-enum34.noarch 0:1.0.4-1.el7    \n  python2-blivet3.noarch 1:3.1.3-3.el7  \n\nDependency Installed:\n  blivet3-data.noarch 1:3.1.3-3.el7                                             \n  device-mapper-event.x86_64 7:1.02.170-6.el7_9.5                               \n  device-mapper-event-libs.x86_64 7:1.02.170-6.el7_9.5                          \n  device-mapper-persistent-data.x86_64 0:0.8.5-3.el7_9.2                        \n  dmraid.x86_64 0:1.0.0.rc16-28.el7                                             \n  dmraid-events.x86_64 0:1.0.0.rc16-28.el7                                      \n  json-glib.x86_64 0:1.4.2-2.el7                                                \n  libaio.x86_64 0:0.3.109-13.el7                                                \n  libblockdev.x86_64 0:2.18-5.el7                                               \n  libblockdev-utils.x86_64 0:2.18-5.el7                                         \n  libbytesize.x86_64 0:1.2-1.el7                                                \n  libdnf.x86_64 0:0.22.5-2.el7_9                                                \n  libmodulemd.x86_64 0:1.6.3-1.el7                                              \n  librepo.x86_64 0:1.8.1-8.el7_9                                                \n  libreport-filesystem.x86_64 0:2.1.11-53.el7                                   \n  librhsm.x86_64 0:0.0.3-3.el7_9                                                \n  libsolv.x86_64 0:0.6.34-4.el7                                                 \n  lsof.x86_64 0:4.87-6.el7                                                      \n  lvm2.x86_64 7:2.02.187-6.el7_9.5                                              \n  lvm2-libs.x86_64 7:2.02.187-6.el7_9.5                                         \n  mdadm.x86_64 0:4.1-9.el7_9                                                    \n  mpfr.x86_64 0:3.1.1-4.el7                                                     \n  pyparted.x86_64 1:3.9-15.el7                                                  \n  python2-blockdev.x86_64 0:2.18-5.el7                                          \n  python2-bytesize.x86_64 0:1.2-1.el7                                           \n  python2-hawkey.x86_64 0:0.22.5-2.el7_9                                        \n  python2-libdnf.x86_64 0:0.22.5-2.el7_9                                        \n  sgpio.x86_64 0:1.2.0.10-13.el7                                                \n  volume_key-libs.x86_64 0:0.3.9-9.el7                                          \n\nComplete!\n"
    ]
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:52:41 +0000 (0:00:08.957)       0:00:12.932 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:52:41 +0000 (0:00:00.037)       0:00:12.970 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:52:41 +0000 (0:00:00.071)       0:00:13.041 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:52:42 +0000 (0:00:00.640)       0:00:13.682 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Thursday 21 July 2022  18:52:42 +0000 (0:00:00.041)       0:00:13.723 ********* 

TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Thursday 21 July 2022  18:52:42 +0000 (0:00:00.065)       0:00:13.788 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable COPRs] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18
Thursday 21 July 2022  18:52:42 +0000 (0:00:00.037)       0:00:13.825 ********* 

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:52:42 +0000 (0:00:00.031)       0:00:13.857 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "rc": 0,
    "results": []
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:52:43 +0000 (0:00:00.548)       0:00:14.406 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "arp-ethers.service": {
                "name": "arp-ethers.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "brandbot.service": {
                "name": "brandbot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "console-shell.service": {
                "name": "console-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "display-manager.service": {
                "name": "display-manager.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dmraid-activation.service": {
                "name": "dmraid-activation.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "exim.service": {
                "name": "exim.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "active"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ip6tables.service": {
                "name": "ip6tables.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "iptables.service": {
                "name": "iptables.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-activation-early.service": {
                "name": "lvm2-activation-early.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "lvm2-activation.service": {
                "name": "lvm2-activation.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "netconsole": {
                "name": "netconsole",
                "source": "sysv",
                "state": "stopped",
                "status": "disabled"
            },
            "network": {
                "name": "network",
                "source": "sysv",
                "state": "running",
                "status": "enabled"
            },
            "network.service": {
                "name": "network.service",
                "source": "systemd",
                "state": "stopped",
                "status": "active"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-config.service": {
                "name": "nfs-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-idmap.service": {
                "name": "nfs-idmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-lock.service": {
                "name": "nfs-lock.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-secure-server.service": {
                "name": "nfs-secure-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "nfs-secure.service": {
                "name": "nfs-secure.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs.service": {
                "name": "nfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfslock.service": {
                "name": "nfslock.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "ntpd.service": {
                "name": "ntpd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "ntpdate.service": {
                "name": "ntpdate.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "ovirt-guest-agent.service": {
                "name": "ovirt-guest-agent.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "postfix.service": {
                "name": "postfix.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rhcd.service": {
                "name": "rhcd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-configure.service": {
                "name": "rhel-configure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-domainname.service": {
                "name": "rhel-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-import-state.service": {
                "name": "rhel-import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-readonly.service": {
                "name": "rhel-readonly.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhnsd": {
                "name": "rhnsd",
                "source": "sysv",
                "state": "running",
                "status": "enabled"
            },
            "rhnsd.service": {
                "name": "rhnsd.service",
                "source": "systemd",
                "state": "running",
                "status": "active"
            },
            "rhsm-facts.service": {
                "name": "rhsm-facts.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rhsm.service": {
                "name": "rhsm.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rhsmcertd.service": {
                "name": "rhsmcertd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpcgssd.service": {
                "name": "rpcgssd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rpcidmapd.service": {
                "name": "rpcidmapd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rsyncd.service": {
                "name": "rsyncd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rsyncd@.service": {
                "name": "rsyncd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service",
                "source": "systemd",
                "state": "stopped",
                "status": "active"
            },
            "sendmail.service": {
                "name": "sendmail.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service",
                "source": "systemd",
                "state": "running",
                "status": "active"
            },
            "sntp.service": {
                "name": "sntp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "sshd-keygen.service": {
                "name": "sshd-keygen.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-importd.service": {
                "name": "systemd-importd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-machined.service": {
                "name": "systemd-machined.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-timesyncd.service": {
                "name": "systemd-timesyncd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ypbind.service": {
                "name": "ypbind.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "yppasswdd.service": {
                "name": "yppasswdd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "ypserv.service": {
                "name": "ypserv.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "ypxfrd.service": {
                "name": "ypxfrd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:52:44 +0000 (0:00:01.140)       0:00:15.546 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:52:44 +0000 (0:00:00.056)       0:00:15.603 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:52:44 +0000 (0:00:00.021)       0:00:15.624 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78
Thursday 21 July 2022  18:52:44 +0000 (0:00:00.484)       0:00:16.109 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:52:44 +0000 (0:00:00.035)       0:00:16.145 ********* 

TASK [fedora.linux_system_roles.storage : show blivet_output] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96
Thursday 21 July 2022  18:52:44 +0000 (0:00:00.020)       0:00:16.165 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [],
        "changed": false,
        "crypts": [],
        "failed": false,
        "leaves": [],
        "mounts": [],
        "packages": [],
        "pools": [],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101
Thursday 21 July 2022  18:52:44 +0000 (0:00:00.038)       0:00:16.203 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105
Thursday 21 July 2022  18:52:44 +0000 (0:00:00.036)       0:00:16.240 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : remove obsolete mounts] **************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121
Thursday 21 July 2022  18:52:44 +0000 (0:00:00.036)       0:00:16.277 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Thursday 21 July 2022  18:52:44 +0000 (0:00:00.031)       0:00:16.309 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : set up new/current mounts] ***********
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137
Thursday 21 July 2022  18:52:44 +0000 (0:00:00.021)       0:00:16.330 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Thursday 21 July 2022  18:52:45 +0000 (0:00:00.030)       0:00:16.361 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156
Thursday 21 July 2022  18:52:45 +0000 (0:00:00.021)       0:00:16.382 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429559.3586438,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161
Thursday 21 July 2022  18:52:45 +0000 (0:00:00.430)       0:00:16.813 ********* 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183
Thursday 21 July 2022  18:52:45 +0000 (0:00:00.023)       0:00:16.836 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [Mark tasks to be skipped] ************************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:18
Thursday 21 July 2022  18:52:46 +0000 (0:00:00.904)       0:00:17.740 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_skip_checks": [
            "blivet_available",
            "packages_installed",
            "service_facts"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:25
Thursday 21 July 2022  18:52:46 +0000 (0:00:00.076)       0:00:17.816 ********* 
included: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml for /cache/rhel-7.qcow2

TASK [Find unused disks in the system] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml:2
Thursday 21 July 2022  18:52:46 +0000 (0:00:00.063)       0:00:17.880 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "disks": [
        "sdb"
    ]
}

TASK [Set unused_disks if necessary] *******************************************
task path: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml:9
Thursday 21 July 2022  18:52:46 +0000 (0:00:00.458)       0:00:18.339 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "unused_disks": [
            "sdb"
        ]
    },
    "changed": false
}

TASK [Exit playbook when there's not enough unused disks in the system] ********
task path: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml:14
Thursday 21 July 2022  18:52:47 +0000 (0:00:00.037)       0:00:18.377 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Print unused disks] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml:19
Thursday 21 July 2022  18:52:47 +0000 (0:00:00.036)       0:00:18.413 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "unused_disks": [
        "sdb"
    ]
}

TASK [Test creating ext4 filesystem with valid parameter "-Fb 4096"] ***********
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:30
Thursday 21 July 2022  18:52:47 +0000 (0:00:00.034)       0:00:18.448 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:52:47 +0000 (0:00:00.043)       0:00:18.491 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:52:47 +0000 (0:00:00.034)       0:00:18.525 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:52:47 +0000 (0:00:00.415)       0:00:18.941 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:52:47 +0000 (0:00:00.059)       0:00:19.001 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:52:47 +0000 (0:00:00.032)       0:00:19.033 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:52:47 +0000 (0:00:00.035)       0:00:19.069 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:52:47 +0000 (0:00:00.055)       0:00:19.125 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:52:47 +0000 (0:00:00.023)       0:00:19.148 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:52:47 +0000 (0:00:00.031)       0:00:19.179 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "sdb"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_create_options": "-Fb 4096",
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:52:47 +0000 (0:00:00.036)       0:00:19.216 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:52:47 +0000 (0:00:00.036)       0:00:19.252 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:52:47 +0000 (0:00:00.028)       0:00:19.281 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:52:47 +0000 (0:00:00.026)       0:00:19.308 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:52:47 +0000 (0:00:00.026)       0:00:19.334 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:52:48 +0000 (0:00:00.027)       0:00:19.361 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:52:48 +0000 (0:00:00.067)       0:00:19.428 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:52:48 +0000 (0:00:00.023)       0:00:19.452 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "create format",
            "device": "/dev/sdb",
            "fs_type": "lvmpv"
        },
        {
            "action": "create device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "create device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd",
        "/dev/mapper/foo-test1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext4",
            "opts": "defaults",
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs",
        "lvm2"
    ],
    "pools": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "-Fb 4096",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78
Thursday 21 July 2022  18:52:49 +0000 (0:00:01.734)       0:00:21.186 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:52:49 +0000 (0:00:00.072)       0:00:21.258 ********* 

TASK [fedora.linux_system_roles.storage : show blivet_output] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96
Thursday 21 July 2022  18:52:49 +0000 (0:00:00.022)       0:00:21.281 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "create format",
                "device": "/dev/sdb",
                "fs_type": "lvmpv"
            },
            {
                "action": "create device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "create device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd",
            "/dev/mapper/foo-test1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext4",
                "opts": "defaults",
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs",
            "lvm2"
        ],
        "pools": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "-Fb 4096",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101
Thursday 21 July 2022  18:52:49 +0000 (0:00:00.045)       0:00:21.326 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "-Fb 4096",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105
Thursday 21 July 2022  18:52:50 +0000 (0:00:00.041)       0:00:21.367 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : remove obsolete mounts] **************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121
Thursday 21 July 2022  18:52:50 +0000 (0:00:00.037)       0:00:21.405 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Thursday 21 July 2022  18:52:50 +0000 (0:00:00.039)       0:00:21.444 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : set up new/current mounts] ***********
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137
Thursday 21 July 2022  18:52:50 +0000 (0:00:00.727)       0:00:22.171 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'ext4', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "opts": "defaults",
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Thursday 21 July 2022  18:52:51 +0000 (0:00:00.483)       0:00:22.655 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156
Thursday 21 July 2022  18:52:51 +0000 (0:00:00.470)       0:00:23.125 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429559.3586438,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161
Thursday 21 July 2022  18:52:52 +0000 (0:00:00.321)       0:00:23.446 ********* 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183
Thursday 21 July 2022  18:52:52 +0000 (0:00:00.022)       0:00:23.469 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:45
Thursday 21 July 2022  18:52:52 +0000 (0:00:00.847)       0:00:24.316 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1
Thursday 21 July 2022  18:52:53 +0000 (0:00:00.037)       0:00:24.354 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "-Fb 4096",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6
Thursday 21 July 2022  18:52:53 +0000 (0:00:00.051)       0:00:24.406 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14
Thursday 21 July 2022  18:52:53 +0000 (0:00:00.037)       0:00:24.443 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/mapper/foo-test1": {
            "fstype": "ext4",
            "label": "",
            "name": "/dev/mapper/foo-test1",
            "size": "4G",
            "type": "lvm",
            "uuid": "4816481a-a158-4c18-a9ff-9e8ad37f71ba"
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "LVM2_member",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": "1XQT5V-vuVQ-F1kB-I2b3-OwaT-XCqy-1rhcZ9"
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-18-52-16-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19
Thursday 21 July 2022  18:52:53 +0000 (0:00:00.433)       0:00:24.877 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003658",
    "end": "2022-07-21 14:52:53.939596",
    "rc": 0,
    "start": "2022-07-21 14:52:53.935938"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0
/dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24
Thursday 21 July 2022  18:52:53 +0000 (0:00:00.429)       0:00:25.307 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003457",
    "end": "2022-07-21 14:52:54.258008",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 14:52:54.254551"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33
Thursday 21 July 2022  18:52:54 +0000 (0:00:00.320)       0:00:25.627 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None})

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:5
Thursday 21 July 2022  18:52:54 +0000 (0:00:00.068)       0:00:25.695 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:18
Thursday 21 July 2022  18:52:54 +0000 (0:00:00.034)       0:00:25.730 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members)
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes)

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:1
Thursday 21 July 2022  18:52:54 +0000 (0:00:00.046)       0:00:25.776 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sdb"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:6
Thursday 21 July 2022  18:52:54 +0000 (0:00:00.052)       0:00:25.828 ********* 
ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sdb",
    "pv": "/dev/sdb"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:15
Thursday 21 July 2022  18:52:54 +0000 (0:00:00.443)       0:00:26.272 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:19
Thursday 21 July 2022  18:52:54 +0000 (0:00:00.050)       0:00:26.322 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sdb"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:23
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.050)       0:00:26.373 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:29
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.051)       0:00:26.424 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:33
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.039)       0:00:26.463 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:37
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.047)       0:00:26.511 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:41
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.022)       0:00:26.534 ********* 
ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sdb"
}

MSG:

All assertions passed

TASK [Check MD RAID] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:50
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.041)       0:00:26.576 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml for /cache/rhel-7.qcow2

TASK [get information about RAID] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:6
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.040)       0:00:26.616 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:12
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.022)       0:00:26.639 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:16
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.022)       0:00:26.662 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:20
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.023)       0:00:26.685 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:24
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.025)       0:00:26.710 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:30
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.024)       0:00:26.734 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:36
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.026)       0:00:26.760 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:44
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.024)       0:00:26.785 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:53
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.068)       0:00:26.853 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml:1
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.076)       0:00:26.930 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'})

TASK [Get information about LVM RAID] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.042)       0:00:26.972 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is LVM RAID] *******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.026)       0:00:26.998 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.029)       0:00:27.028 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:56
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.027)       0:00:27.056 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml for /cache/rhel-7.qcow2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml:1
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.042)       0:00:27.098 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'})

TASK [Get information about thinpool] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.044)       0:00:27.142 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.025)       0:00:27.168 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.025)       0:00:27.193 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.025)       0:00:27.219 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check member encryption] *************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:59
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.024)       0:00:27.244 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:4
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.047)       0:00:27.291 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:8
Thursday 21 July 2022  18:52:55 +0000 (0:00:00.049)       0:00:27.341 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=/dev/sdb)  => {
    "_storage_test_pool_member_path": "/dev/sdb",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:15
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.026)       0:00:27.368 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/sdb)

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:1
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.040)       0:00:27.408 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:4
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.047)       0:00:27.456 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:9
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.045)       0:00:27.502 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:15
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.033)       0:00:27.535 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:21
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.036)       0:00:27.571 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:27
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.033)       0:00:27.604 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:22
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.034)       0:00:27.639 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:62
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.033)       0:00:27.673 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml:1
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.049)       0:00:27.722 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'})

TASK [get information about VDO deduplication] *********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.043)       0:00:27.766 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.031)       0:00:27.797 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.026)       0:00:27.824 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.023)       0:00:27.848 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.025)       0:00:27.874 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.025)       0:00:27.899 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.026)       0:00:27.926 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.024)       0:00:27.950 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:65
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.036)       0:00:27.987 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [verify the volumes] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml:3
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.036)       0:00:28.024 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'})

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.041)       0:00:28.065 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.122)       0:00:28.188 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache)

TASK [Get expected mount device based on device type] **************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.082)       0:00:28.271 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10
Thursday 21 July 2022  18:52:56 +0000 (0:00:00.042)       0:00:28.313 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": [
            {
                "block_available": 938726,
                "block_size": 4096,
                "block_total": 999344,
                "block_used": 60618,
                "device": "/dev/mapper/foo-test1",
                "fstype": "ext4",
                "inode_available": 262133,
                "inode_total": 262144,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 3845021696,
                "size_total": 4093313024,
                "uuid": "4816481a-a158-4c18-a9ff-9e8ad37f71ba"
            }
        ],
        "storage_test_mount_expected_match_count": "1",
        "storage_test_mount_point_matches": [
            {
                "block_available": 938726,
                "block_size": 4096,
                "block_total": 999344,
                "block_used": 60618,
                "device": "/dev/mapper/foo-test1",
                "fstype": "ext4",
                "inode_available": 262133,
                "inode_total": 262144,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 3845021696,
                "size_total": 4093313024,
                "uuid": "4816481a-a158-4c18-a9ff-9e8ad37f71ba"
            }
        ],
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20
Thursday 21 July 2022  18:52:57 +0000 (0:00:00.056)       0:00:28.369 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the current mount state by mount point] ***************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29
Thursday 21 July 2022  18:52:57 +0000 (0:00:00.051)       0:00:28.421 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the mount fs type] ************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37
Thursday 21 July 2022  18:52:57 +0000 (0:00:00.047)       0:00:28.468 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [command] *****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46
Thursday 21 July 2022  18:52:57 +0000 (0:00:00.051)       0:00:28.520 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50
Thursday 21 July 2022  18:52:57 +0000 (0:00:00.025)       0:00:28.545 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55
Thursday 21 July 2022  18:52:57 +0000 (0:00:00.024)       0:00:28.570 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65
Thursday 21 July 2022  18:52:57 +0000 (0:00:00.024)       0:00:28.594 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": null,
        "storage_test_mount_expected_match_count": null,
        "storage_test_mount_point_matches": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2
Thursday 21 July 2022  18:52:57 +0000 (0:00:00.035)       0:00:28.630 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12
Thursday 21 July 2022  18:52:57 +0000 (0:00:00.061)       0:00:28.691 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19
Thursday 21 July 2022  18:52:57 +0000 (0:00:00.051)       0:00:28.743 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25
Thursday 21 July 2022  18:52:57 +0000 (0:00:00.049)       0:00:28.792 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up variables] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34
Thursday 21 July 2022  18:52:57 +0000 (0:00:00.039)       0:00:28.832 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4
Thursday 21 July 2022  18:52:57 +0000 (0:00:00.034)       0:00:28.866 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10
Thursday 21 July 2022  18:52:57 +0000 (0:00:00.039)       0:00:28.905 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4
Thursday 21 July 2022  18:52:57 +0000 (0:00:00.039)       0:00:28.945 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429569.7756436,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1658429569.7756436,
        "dev": 5,
        "device_type": 64512,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 32358,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1658429569.7756436,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10
Thursday 21 July 2022  18:52:57 +0000 (0:00:00.322)       0:00:29.267 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15
Thursday 21 July 2022  18:52:57 +0000 (0:00:00.039)       0:00:29.307 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [(1/2) Process volume type (set initial value)] ***************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21
Thursday 21 July 2022  18:52:57 +0000 (0:00:00.040)       0:00:29.347 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [(2/2) Process volume type (get RAID value)] ******************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25
Thursday 21 July 2022  18:52:58 +0000 (0:00:00.038)       0:00:29.386 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30
Thursday 21 July 2022  18:52:58 +0000 (0:00:00.025)       0:00:29.412 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3
Thursday 21 July 2022  18:52:58 +0000 (0:00:00.038)       0:00:29.450 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10
Thursday 21 July 2022  18:52:58 +0000 (0:00:00.024)       0:00:29.475 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "changed": true,
    "changes": {
        "installed": [
            "cryptsetup"
        ]
    },
    "rc": 0,
    "results": [
        "Loaded plugins: search-disabled-repos\nResolving Dependencies\n--> Running transaction check\n---> Package cryptsetup.x86_64 0:2.0.3-6.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package             Arch            Version                Repository     Size\n================================================================================\nInstalling:\n cryptsetup          x86_64          2.0.3-6.el7            rhel          154 k\n\nTransaction Summary\n================================================================================\nInstall  1 Package\n\nTotal download size: 154 k\nInstalled size: 354 k\nDownloading packages:\nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n  Installing : cryptsetup-2.0.3-6.el7.x86_64                                1/1 \n  Verifying  : cryptsetup-2.0.3-6.el7.x86_64                                1/1 \n\nInstalled:\n  cryptsetup.x86_64 0:2.0.3-6.el7                                               \n\nComplete!\n"
    ]
}

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15
Thursday 21 July 2022  18:52:59 +0000 (0:00:01.620)       0:00:31.096 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21
Thursday 21 July 2022  18:52:59 +0000 (0:00:00.023)       0:00:31.119 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27
Thursday 21 July 2022  18:52:59 +0000 (0:00:00.024)       0:00:31.144 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33
Thursday 21 July 2022  18:52:59 +0000 (0:00:00.081)       0:00:31.225 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39
Thursday 21 July 2022  18:52:59 +0000 (0:00:00.025)       0:00:31.251 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44
Thursday 21 July 2022  18:52:59 +0000 (0:00:00.022)       0:00:31.273 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50
Thursday 21 July 2022  18:52:59 +0000 (0:00:00.022)       0:00:31.296 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56
Thursday 21 July 2022  18:52:59 +0000 (0:00:00.021)       0:00:31.318 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62
Thursday 21 July 2022  18:52:59 +0000 (0:00:00.026)       0:00:31.344 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67
Thursday 21 July 2022  18:53:00 +0000 (0:00:00.078)       0:00:31.423 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72
Thursday 21 July 2022  18:53:00 +0000 (0:00:00.050)       0:00:31.473 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78
Thursday 21 July 2022  18:53:00 +0000 (0:00:00.037)       0:00:31.511 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84
Thursday 21 July 2022  18:53:00 +0000 (0:00:00.045)       0:00:31.557 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90
Thursday 21 July 2022  18:53:00 +0000 (0:00:00.040)       0:00:31.597 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [get information about RAID] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7
Thursday 21 July 2022  18:53:00 +0000 (0:00:00.033)       0:00:31.631 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13
Thursday 21 July 2022  18:53:00 +0000 (0:00:00.038)       0:00:31.670 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17
Thursday 21 July 2022  18:53:00 +0000 (0:00:00.043)       0:00:31.713 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21
Thursday 21 July 2022  18:53:00 +0000 (0:00:00.038)       0:00:31.751 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25
Thursday 21 July 2022  18:53:00 +0000 (0:00:00.038)       0:00:31.790 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31
Thursday 21 July 2022  18:53:00 +0000 (0:00:00.037)       0:00:31.828 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37
Thursday 21 July 2022  18:53:00 +0000 (0:00:00.039)       0:00:31.868 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the actual size of the volume] *************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3
Thursday 21 July 2022  18:53:00 +0000 (0:00:00.036)       0:00:31.905 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [parse the requested size of the volume] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9
Thursday 21 July 2022  18:53:00 +0000 (0:00:00.415)       0:00:32.320 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15
Thursday 21 July 2022  18:53:01 +0000 (0:00:00.326)       0:00:32.646 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_expected_size": "4294967296"
    },
    "changed": false
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20
Thursday 21 July 2022  18:53:01 +0000 (0:00:00.050)       0:00:32.697 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25
Thursday 21 July 2022  18:53:01 +0000 (0:00:00.036)       0:00:32.734 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28
Thursday 21 July 2022  18:53:01 +0000 (0:00:00.034)       0:00:32.768 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31
Thursday 21 July 2022  18:53:01 +0000 (0:00:00.036)       0:00:32.805 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36
Thursday 21 July 2022  18:53:01 +0000 (0:00:00.033)       0:00:32.838 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39
Thursday 21 July 2022  18:53:01 +0000 (0:00:00.034)       0:00:32.872 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44
Thursday 21 July 2022  18:53:01 +0000 (0:00:00.035)       0:00:32.908 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_actual_size": {
        "bytes": 4294967296,
        "changed": false,
        "failed": false,
        "lvm": "4g",
        "parted": "4GiB",
        "size": "4 GiB"
    }
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47
Thursday 21 July 2022  18:53:01 +0000 (0:00:00.039)       0:00:32.947 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50
Thursday 21 July 2022  18:53:01 +0000 (0:00:00.036)       0:00:32.983 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6
Thursday 21 July 2022  18:53:01 +0000 (0:00:00.054)       0:00:33.037 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.028349",
    "end": "2022-07-21 14:53:02.018909",
    "rc": 0,
    "start": "2022-07-21 14:53:01.990560"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14
Thursday 21 July 2022  18:53:02 +0000 (0:00:00.350)       0:00:33.388 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [check segment type] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17
Thursday 21 July 2022  18:53:02 +0000 (0:00:00.052)       0:00:33.440 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22
Thursday 21 July 2022  18:53:02 +0000 (0:00:00.050)       0:00:33.490 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested cache size] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26
Thursday 21 July 2022  18:53:02 +0000 (0:00:00.037)       0:00:33.527 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32
Thursday 21 July 2022  18:53:02 +0000 (0:00:00.038)       0:00:33.566 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36
Thursday 21 July 2022  18:53:02 +0000 (0:00:00.037)       0:00:33.604 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16
Thursday 21 July 2022  18:53:02 +0000 (0:00:00.037)       0:00:33.642 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43
Thursday 21 July 2022  18:53:02 +0000 (0:00:00.072)       0:00:33.715 ********* 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53
Thursday 21 July 2022  18:53:02 +0000 (0:00:00.022)       0:00:33.737 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Remove the volume group created above] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:47
Thursday 21 July 2022  18:53:02 +0000 (0:00:00.073)       0:00:33.810 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:53:02 +0000 (0:00:00.082)       0:00:33.893 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:53:02 +0000 (0:00:00.036)       0:00:33.929 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:53:03 +0000 (0:00:00.422)       0:00:34.352 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:53:03 +0000 (0:00:00.061)       0:00:34.413 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:53:03 +0000 (0:00:00.038)       0:00:34.452 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:53:03 +0000 (0:00:00.036)       0:00:34.488 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:53:03 +0000 (0:00:00.056)       0:00:34.545 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:53:03 +0000 (0:00:00.023)       0:00:34.568 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:53:03 +0000 (0:00:00.031)       0:00:34.600 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "sdb"
            ],
            "name": "foo",
            "state": "absent",
            "type": "lvm"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:53:03 +0000 (0:00:00.041)       0:00:34.642 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:53:03 +0000 (0:00:00.039)       0:00:34.681 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:53:03 +0000 (0:00:00.031)       0:00:34.712 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:53:03 +0000 (0:00:00.028)       0:00:34.741 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:53:03 +0000 (0:00:00.028)       0:00:34.769 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:53:03 +0000 (0:00:00.028)       0:00:34.798 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:53:03 +0000 (0:00:00.057)       0:00:34.855 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:53:03 +0000 (0:00:00.023)       0:00:34.879 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "destroy device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/sdb",
            "fs_type": "lvmpv"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [
        {
            "fstype": "ext4",
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "absent"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "absent",
            "type": "lvm",
            "volumes": []
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78
Thursday 21 July 2022  18:53:05 +0000 (0:00:01.650)       0:00:36.530 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:53:05 +0000 (0:00:00.039)       0:00:36.570 ********* 

TASK [fedora.linux_system_roles.storage : show blivet_output] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96
Thursday 21 July 2022  18:53:05 +0000 (0:00:00.024)       0:00:36.594 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "destroy device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/sdb",
                "fs_type": "lvmpv"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [
            {
                "fstype": "ext4",
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "absent"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101
Thursday 21 July 2022  18:53:05 +0000 (0:00:00.076)       0:00:36.671 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105
Thursday 21 July 2022  18:53:05 +0000 (0:00:00.040)       0:00:36.711 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : remove obsolete mounts] **************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121
Thursday 21 July 2022  18:53:05 +0000 (0:00:00.037)       0:00:36.749 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'ext4'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "fstype": "ext4",
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Thursday 21 July 2022  18:53:05 +0000 (0:00:00.383)       0:00:37.132 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : set up new/current mounts] ***********
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137
Thursday 21 July 2022  18:53:06 +0000 (0:00:00.529)       0:00:37.662 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Thursday 21 July 2022  18:53:06 +0000 (0:00:00.040)       0:00:37.702 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156
Thursday 21 July 2022  18:53:06 +0000 (0:00:00.555)       0:00:38.257 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429559.3586438,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161
Thursday 21 July 2022  18:53:07 +0000 (0:00:00.356)       0:00:38.613 ********* 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183
Thursday 21 July 2022  18:53:07 +0000 (0:00:00.023)       0:00:38.637 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:57
Thursday 21 July 2022  18:53:08 +0000 (0:00:00.837)       0:00:39.474 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1
Thursday 21 July 2022  18:53:08 +0000 (0:00:00.041)       0:00:39.516 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "absent",
            "type": "lvm",
            "volumes": []
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6
Thursday 21 July 2022  18:53:08 +0000 (0:00:00.047)       0:00:39.563 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14
Thursday 21 July 2022  18:53:08 +0000 (0:00:00.033)       0:00:39.597 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-18-52-16-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19
Thursday 21 July 2022  18:53:08 +0000 (0:00:00.317)       0:00:39.914 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003655",
    "end": "2022-07-21 14:53:08.859192",
    "rc": 0,
    "start": "2022-07-21 14:53:08.855537"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24
Thursday 21 July 2022  18:53:08 +0000 (0:00:00.314)       0:00:40.229 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003771",
    "end": "2022-07-21 14:53:09.180798",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 14:53:09.177027"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33
Thursday 21 July 2022  18:53:09 +0000 (0:00:00.323)       0:00:40.552 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'absent', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None})

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:5
Thursday 21 July 2022  18:53:09 +0000 (0:00:00.056)       0:00:40.609 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:18
Thursday 21 July 2022  18:53:09 +0000 (0:00:00.033)       0:00:40.643 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members)
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes)

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:1
Thursday 21 July 2022  18:53:09 +0000 (0:00:00.046)       0:00:40.689 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "0",
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:6
Thursday 21 July 2022  18:53:09 +0000 (0:00:00.083)       0:00:40.772 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:15
Thursday 21 July 2022  18:53:09 +0000 (0:00:00.022)       0:00:40.795 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": "0"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:19
Thursday 21 July 2022  18:53:09 +0000 (0:00:00.049)       0:00:40.845 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": []
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:23
Thursday 21 July 2022  18:53:09 +0000 (0:00:00.083)       0:00:40.928 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:29
Thursday 21 July 2022  18:53:09 +0000 (0:00:00.088)       0:00:41.016 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:33
Thursday 21 July 2022  18:53:09 +0000 (0:00:00.039)       0:00:41.055 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:37
Thursday 21 July 2022  18:53:09 +0000 (0:00:00.081)       0:00:41.137 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:41
Thursday 21 July 2022  18:53:09 +0000 (0:00:00.025)       0:00:41.162 ********* 

TASK [Check MD RAID] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:50
Thursday 21 July 2022  18:53:09 +0000 (0:00:00.047)       0:00:41.209 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml for /cache/rhel-7.qcow2

TASK [get information about RAID] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:6
Thursday 21 July 2022  18:53:09 +0000 (0:00:00.042)       0:00:41.252 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:12
Thursday 21 July 2022  18:53:09 +0000 (0:00:00.024)       0:00:41.276 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:16
Thursday 21 July 2022  18:53:09 +0000 (0:00:00.026)       0:00:41.303 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:20
Thursday 21 July 2022  18:53:09 +0000 (0:00:00.026)       0:00:41.329 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:24
Thursday 21 July 2022  18:53:09 +0000 (0:00:00.022)       0:00:41.352 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:30
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.025)       0:00:41.377 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:36
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.022)       0:00:41.400 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:44
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.024)       0:00:41.424 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:53
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.047)       0:00:41.471 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml:1
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.049)       0:00:41.521 ********* 

TASK [Check Thin Pools] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:56
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.021)       0:00:41.542 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml for /cache/rhel-7.qcow2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml:1
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.043)       0:00:41.585 ********* 

TASK [Check member encryption] *************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:59
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.021)       0:00:41.606 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:4
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.044)       0:00:41.650 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:8
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.048)       0:00:41.699 ********* 

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:15
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.020)       0:00:41.719 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:22
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.019)       0:00:41.739 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:62
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.033)       0:00:41.773 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml:1
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.049)       0:00:41.822 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:65
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.023)       0:00:41.845 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [verify the volumes] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml:3
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.037)       0:00:41.883 ********* 

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.021)       0:00:41.905 ********* 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.021)       0:00:41.927 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Try to create ext4 filesystem with invalid parameter "-Fb 512"] **********
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:61
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.034)       0:00:41.961 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.041)       0:00:42.003 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:53:10 +0000 (0:00:00.036)       0:00:42.039 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:53:11 +0000 (0:00:00.420)       0:00:42.460 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:53:11 +0000 (0:00:00.095)       0:00:42.556 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:53:11 +0000 (0:00:00.100)       0:00:42.656 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:53:11 +0000 (0:00:00.034)       0:00:42.690 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:53:11 +0000 (0:00:00.054)       0:00:42.745 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:53:11 +0000 (0:00:00.019)       0:00:42.764 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:53:11 +0000 (0:00:00.029)       0:00:42.794 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "sdb"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_create_options": "-Fb 512",
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:53:11 +0000 (0:00:00.037)       0:00:42.832 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:53:11 +0000 (0:00:00.036)       0:00:42.868 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:53:11 +0000 (0:00:00.026)       0:00:42.895 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:53:11 +0000 (0:00:00.027)       0:00:42.923 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:53:11 +0000 (0:00:00.027)       0:00:42.950 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:53:11 +0000 (0:00:00.034)       0:00:42.985 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:53:11 +0000 (0:00:00.060)       0:00:43.045 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:53:11 +0000 (0:00:00.023)       0:00:43.068 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "xfsprogs",
        "e2fsprogs",
        "lvm2"
    ],
    "pools": [],
    "volumes": []
}

MSG:

Failed to commit changes to disk: (FSError('format failed: 1',), '/dev/mapper/foo-test1')

TASK [fedora.linux_system_roles.storage : failed message] **********************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:86
Thursday 21 July 2022  18:53:13 +0000 (0:00:01.298)       0:00:44.367 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "changed": false
}

MSG:

{'crypts': [], 'mounts': [], 'leaves': [], 'changed': False, 'actions': [], 'failed': True, 'volumes': [], 'invocation': {'module_args': {'packages_only': False, 'disklabel_type': None, 'diskvolume_mkfs_option_map': {'ext4': '-F', 'ext3': '-F', 'ext2': '-F'}, 'safe_mode': False, 'pools': [{'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 512'}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}], 'volumes': [], 'pool_defaults': {'encryption_password': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_cipher': None, 'disks': [], 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_device_count': None, 'state': 'present', 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'raid_level': None, 'raid_spare_count': None}, 'volume_defaults': {'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'xfs', 'mount_options': 'defaults', 'size': 0, 'mount_point': '', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'fs_overwrite_existing': True, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'mount_passno': 0, 'raid_spare_count': None, 'cache_mode': None, 'deduplication': None, 'cached': False, 'type': 'lvm', 'disks': [], 'thin_pool_size': None, 'thin': None, 'mount_check': 0, 'cache_size': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, 'use_partitions': None}}, 'pools': [], 'packages': ['xfsprogs', 'e2fsprogs', 'lvm2'], 'msg': "Failed to commit changes to disk: (FSError('format failed: 1',), '/dev/mapper/foo-test1')", '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:53:13 +0000 (0:00:00.040)       0:00:44.407 ********* 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:81
Thursday 21 July 2022  18:53:13 +0000 (0:00:00.021)       0:00:44.429 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the output when creating ext4 filesystem with invalid parameter "-Fb 512"] ***
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:87
Thursday 21 July 2022  18:53:13 +0000 (0:00:00.039)       0:00:44.469 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Remove the volume group created above] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:92
Thursday 21 July 2022  18:53:13 +0000 (0:00:00.059)       0:00:44.528 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:53:13 +0000 (0:00:00.035)       0:00:44.564 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:53:13 +0000 (0:00:00.037)       0:00:44.601 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:53:13 +0000 (0:00:00.415)       0:00:45.016 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:53:13 +0000 (0:00:00.072)       0:00:45.089 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:53:13 +0000 (0:00:00.038)       0:00:45.127 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:53:13 +0000 (0:00:00.037)       0:00:45.165 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:53:13 +0000 (0:00:00.059)       0:00:45.224 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:53:13 +0000 (0:00:00.020)       0:00:45.244 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:53:13 +0000 (0:00:00.027)       0:00:45.272 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "sdb"
            ],
            "name": "foo",
            "state": "absent",
            "type": "lvm"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:53:14 +0000 (0:00:00.097)       0:00:45.369 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:53:14 +0000 (0:00:00.040)       0:00:45.409 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:53:14 +0000 (0:00:00.031)       0:00:45.440 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:53:14 +0000 (0:00:00.029)       0:00:45.470 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:53:14 +0000 (0:00:00.033)       0:00:45.504 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:53:14 +0000 (0:00:00.028)       0:00:45.532 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:53:14 +0000 (0:00:00.101)       0:00:45.633 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:53:14 +0000 (0:00:00.056)       0:00:45.690 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "destroy device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/sdb",
            "fs_type": "lvmpv"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [],
    "packages": [
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "absent",
            "type": "lvm",
            "volumes": []
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78
Thursday 21 July 2022  18:53:15 +0000 (0:00:01.387)       0:00:47.078 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:53:15 +0000 (0:00:00.039)       0:00:47.117 ********* 

TASK [fedora.linux_system_roles.storage : show blivet_output] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96
Thursday 21 July 2022  18:53:15 +0000 (0:00:00.020)       0:00:47.138 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "destroy device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/sdb",
                "fs_type": "lvmpv"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [],
        "packages": [
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101
Thursday 21 July 2022  18:53:15 +0000 (0:00:00.035)       0:00:47.174 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105
Thursday 21 July 2022  18:53:15 +0000 (0:00:00.036)       0:00:47.210 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : remove obsolete mounts] **************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121
Thursday 21 July 2022  18:53:15 +0000 (0:00:00.038)       0:00:47.249 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Thursday 21 July 2022  18:53:15 +0000 (0:00:00.034)       0:00:47.283 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : set up new/current mounts] ***********
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137
Thursday 21 July 2022  18:53:15 +0000 (0:00:00.023)       0:00:47.307 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Thursday 21 July 2022  18:53:15 +0000 (0:00:00.034)       0:00:47.341 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156
Thursday 21 July 2022  18:53:16 +0000 (0:00:00.026)       0:00:47.368 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429559.3586438,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161
Thursday 21 July 2022  18:53:16 +0000 (0:00:00.327)       0:00:47.695 ********* 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183
Thursday 21 July 2022  18:53:16 +0000 (0:00:00.023)       0:00:47.719 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [Create one LVM logical volume with "4g" under one volume group] **********
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:102
Thursday 21 July 2022  18:53:17 +0000 (0:00:00.833)       0:00:48.553 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:53:17 +0000 (0:00:00.042)       0:00:48.596 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:53:17 +0000 (0:00:00.035)       0:00:48.631 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:53:17 +0000 (0:00:00.414)       0:00:49.046 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:53:17 +0000 (0:00:00.063)       0:00:49.109 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:53:17 +0000 (0:00:00.036)       0:00:49.145 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:53:17 +0000 (0:00:00.034)       0:00:49.180 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:53:17 +0000 (0:00:00.052)       0:00:49.233 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:53:17 +0000 (0:00:00.018)       0:00:49.251 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:53:17 +0000 (0:00:00.027)       0:00:49.279 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "sdb"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:53:17 +0000 (0:00:00.034)       0:00:49.313 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:53:18 +0000 (0:00:00.067)       0:00:49.380 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:53:18 +0000 (0:00:00.028)       0:00:49.409 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:53:18 +0000 (0:00:00.072)       0:00:49.482 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:53:18 +0000 (0:00:00.027)       0:00:49.509 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:53:18 +0000 (0:00:00.028)       0:00:49.538 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:53:18 +0000 (0:00:00.075)       0:00:49.613 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:53:18 +0000 (0:00:00.026)       0:00:49.640 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "create format",
            "device": "/dev/sdb",
            "fs_type": "lvmpv"
        },
        {
            "action": "create device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "create device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd",
        "/dev/mapper/foo-test1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext4",
            "opts": "defaults",
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs",
        "lvm2"
    ],
    "pools": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78
Thursday 21 July 2022  18:53:19 +0000 (0:00:01.589)       0:00:51.229 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:53:19 +0000 (0:00:00.040)       0:00:51.269 ********* 

TASK [fedora.linux_system_roles.storage : show blivet_output] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96
Thursday 21 July 2022  18:53:19 +0000 (0:00:00.022)       0:00:51.292 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "create format",
                "device": "/dev/sdb",
                "fs_type": "lvmpv"
            },
            {
                "action": "create device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "create device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd",
            "/dev/mapper/foo-test1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext4",
                "opts": "defaults",
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs",
            "lvm2"
        ],
        "pools": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101
Thursday 21 July 2022  18:53:19 +0000 (0:00:00.044)       0:00:51.337 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105
Thursday 21 July 2022  18:53:20 +0000 (0:00:00.045)       0:00:51.382 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : remove obsolete mounts] **************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121
Thursday 21 July 2022  18:53:20 +0000 (0:00:00.042)       0:00:51.424 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Thursday 21 July 2022  18:53:20 +0000 (0:00:00.040)       0:00:51.464 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : set up new/current mounts] ***********
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137
Thursday 21 July 2022  18:53:20 +0000 (0:00:00.494)       0:00:51.958 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'ext4', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "opts": "defaults",
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Thursday 21 July 2022  18:53:20 +0000 (0:00:00.367)       0:00:52.326 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156
Thursday 21 July 2022  18:53:21 +0000 (0:00:00.452)       0:00:52.778 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429559.3586438,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161
Thursday 21 July 2022  18:53:21 +0000 (0:00:00.328)       0:00:53.106 ********* 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183
Thursday 21 July 2022  18:53:21 +0000 (0:00:00.022)       0:00:53.128 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:116
Thursday 21 July 2022  18:53:22 +0000 (0:00:00.879)       0:00:54.008 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1
Thursday 21 July 2022  18:53:22 +0000 (0:00:00.041)       0:00:54.049 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6
Thursday 21 July 2022  18:53:22 +0000 (0:00:00.090)       0:00:54.139 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14
Thursday 21 July 2022  18:53:22 +0000 (0:00:00.040)       0:00:54.180 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/mapper/foo-test1": {
            "fstype": "ext4",
            "label": "",
            "name": "/dev/mapper/foo-test1",
            "size": "4G",
            "type": "lvm",
            "uuid": "1bcafd54-cae0-4716-93a8-e37d0813bf0d"
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "LVM2_member",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": "VwncIJ-eNkD-ZMrj-P29q-wIbc-fW3h-8gR5kP"
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-18-52-16-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19
Thursday 21 July 2022  18:53:23 +0000 (0:00:00.361)       0:00:54.542 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003416",
    "end": "2022-07-21 14:53:23.480800",
    "rc": 0,
    "start": "2022-07-21 14:53:23.477384"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0
/dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24
Thursday 21 July 2022  18:53:23 +0000 (0:00:00.345)       0:00:54.888 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003625",
    "end": "2022-07-21 14:53:23.838728",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 14:53:23.835103"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33
Thursday 21 July 2022  18:53:23 +0000 (0:00:00.321)       0:00:55.209 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None})

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:5
Thursday 21 July 2022  18:53:23 +0000 (0:00:00.064)       0:00:55.274 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:18
Thursday 21 July 2022  18:53:23 +0000 (0:00:00.035)       0:00:55.309 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members)
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes)

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:1
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.054)       0:00:55.364 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sdb"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:6
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.060)       0:00:55.424 ********* 
ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sdb",
    "pv": "/dev/sdb"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:15
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.330)       0:00:55.755 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:19
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.048)       0:00:55.804 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sdb"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:23
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.049)       0:00:55.853 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:29
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.048)       0:00:55.901 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:33
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.036)       0:00:55.938 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:37
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.052)       0:00:55.990 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:41
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.024)       0:00:56.015 ********* 
ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sdb"
}

MSG:

All assertions passed

TASK [Check MD RAID] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:50
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.041)       0:00:56.056 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml for /cache/rhel-7.qcow2

TASK [get information about RAID] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:6
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.040)       0:00:56.097 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:12
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.022)       0:00:56.119 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:16
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.022)       0:00:56.142 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:20
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.022)       0:00:56.164 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:24
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.023)       0:00:56.188 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:30
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.022)       0:00:56.210 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:36
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.026)       0:00:56.236 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:44
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.026)       0:00:56.263 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:53
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.036)       0:00:56.300 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml:1
Thursday 21 July 2022  18:53:24 +0000 (0:00:00.042)       0:00:56.342 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [Get information about LVM RAID] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.041)       0:00:56.384 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is LVM RAID] *******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.028)       0:00:56.413 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.027)       0:00:56.440 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:56
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.029)       0:00:56.470 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml for /cache/rhel-7.qcow2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml:1
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.092)       0:00:56.562 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [Get information about thinpool] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.044)       0:00:56.607 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.027)       0:00:56.634 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.025)       0:00:56.659 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.022)       0:00:56.682 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check member encryption] *************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:59
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.024)       0:00:56.706 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:4
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.045)       0:00:56.752 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:8
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.048)       0:00:56.801 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=/dev/sdb)  => {
    "_storage_test_pool_member_path": "/dev/sdb",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:15
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.027)       0:00:56.828 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/sdb)

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:1
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.044)       0:00:56.873 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:4
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.051)       0:00:56.924 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:9
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.051)       0:00:56.975 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:15
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.035)       0:00:57.010 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:21
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.034)       0:00:57.044 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:27
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.033)       0:00:57.078 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:22
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.038)       0:00:57.116 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:62
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.035)       0:00:57.152 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml:1
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.048)       0:00:57.201 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [get information about VDO deduplication] *********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.049)       0:00:57.250 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.025)       0:00:57.275 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.024)       0:00:57.299 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.026)       0:00:57.326 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21
Thursday 21 July 2022  18:53:25 +0000 (0:00:00.025)       0:00:57.351 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.025)       0:00:57.376 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.024)       0:00:57.400 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.025)       0:00:57.425 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:65
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.031)       0:00:57.457 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [verify the volumes] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml:3
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.032)       0:00:57.489 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.040)       0:00:57.529 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.045)       0:00:57.575 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache)

TASK [Get expected mount device based on device type] **************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.076)       0:00:57.652 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.084)       0:00:57.736 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": [
            {
                "block_available": 938726,
                "block_size": 4096,
                "block_total": 999344,
                "block_used": 60618,
                "device": "/dev/mapper/foo-test1",
                "fstype": "ext4",
                "inode_available": 262133,
                "inode_total": 262144,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 3845021696,
                "size_total": 4093313024,
                "uuid": "1bcafd54-cae0-4716-93a8-e37d0813bf0d"
            }
        ],
        "storage_test_mount_expected_match_count": "1",
        "storage_test_mount_point_matches": [
            {
                "block_available": 938726,
                "block_size": 4096,
                "block_total": 999344,
                "block_used": 60618,
                "device": "/dev/mapper/foo-test1",
                "fstype": "ext4",
                "inode_available": 262133,
                "inode_total": 262144,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 3845021696,
                "size_total": 4093313024,
                "uuid": "1bcafd54-cae0-4716-93a8-e37d0813bf0d"
            }
        ],
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.130)       0:00:57.867 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the current mount state by mount point] ***************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.051)       0:00:57.919 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the mount fs type] ************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.049)       0:00:57.968 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [command] *****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.050)       0:00:58.018 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.022)       0:00:58.041 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.025)       0:00:58.067 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.028)       0:00:58.096 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": null,
        "storage_test_mount_expected_match_count": null,
        "storage_test_mount_point_matches": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.032)       0:00:58.129 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.062)       0:00:58.191 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.048)       0:00:58.240 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.047)       0:00:58.288 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up variables] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34
Thursday 21 July 2022  18:53:26 +0000 (0:00:00.037)       0:00:58.325 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4
Thursday 21 July 2022  18:53:27 +0000 (0:00:00.037)       0:00:58.362 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10
Thursday 21 July 2022  18:53:27 +0000 (0:00:00.043)       0:00:58.406 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4
Thursday 21 July 2022  18:53:27 +0000 (0:00:00.041)       0:00:58.447 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429599.8126438,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1658429599.8126438,
        "dev": 5,
        "device_type": 64512,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 42685,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1658429599.8126438,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10
Thursday 21 July 2022  18:53:27 +0000 (0:00:00.322)       0:00:58.770 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15
Thursday 21 July 2022  18:53:27 +0000 (0:00:00.040)       0:00:58.810 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [(1/2) Process volume type (set initial value)] ***************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21
Thursday 21 July 2022  18:53:27 +0000 (0:00:00.040)       0:00:58.851 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [(2/2) Process volume type (get RAID value)] ******************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25
Thursday 21 July 2022  18:53:27 +0000 (0:00:00.038)       0:00:58.889 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30
Thursday 21 July 2022  18:53:27 +0000 (0:00:00.025)       0:00:58.914 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3
Thursday 21 July 2022  18:53:27 +0000 (0:00:00.040)       0:00:58.955 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10
Thursday 21 July 2022  18:53:27 +0000 (0:00:00.023)       0:00:58.978 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "rc": 0,
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.552)       0:00:59.530 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.023)       0:00:59.554 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.022)       0:00:59.577 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.053)       0:00:59.630 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.024)       0:00:59.655 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.026)       0:00:59.681 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.025)       0:00:59.707 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.025)       0:00:59.733 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.024)       0:00:59.757 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.050)       0:00:59.808 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.096)       0:00:59.904 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.036)       0:00:59.940 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.036)       0:00:59.977 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.033)       0:01:00.010 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [get information about RAID] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.080)       0:01:00.091 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.038)       0:01:00.129 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.035)       0:01:00.164 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.034)       0:01:00.198 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.033)       0:01:00.232 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.034)       0:01:00.267 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37
Thursday 21 July 2022  18:53:28 +0000 (0:00:00.076)       0:01:00.343 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the actual size of the volume] *************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3
Thursday 21 July 2022  18:53:29 +0000 (0:00:00.035)       0:01:00.379 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [parse the requested size of the volume] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9
Thursday 21 July 2022  18:53:29 +0000 (0:00:00.308)       0:01:00.687 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15
Thursday 21 July 2022  18:53:29 +0000 (0:00:00.325)       0:01:01.013 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_expected_size": "4294967296"
    },
    "changed": false
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20
Thursday 21 July 2022  18:53:29 +0000 (0:00:00.052)       0:01:01.065 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25
Thursday 21 July 2022  18:53:29 +0000 (0:00:00.036)       0:01:01.102 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28
Thursday 21 July 2022  18:53:29 +0000 (0:00:00.037)       0:01:01.139 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31
Thursday 21 July 2022  18:53:29 +0000 (0:00:00.039)       0:01:01.179 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36
Thursday 21 July 2022  18:53:29 +0000 (0:00:00.041)       0:01:01.221 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39
Thursday 21 July 2022  18:53:29 +0000 (0:00:00.039)       0:01:01.260 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44
Thursday 21 July 2022  18:53:29 +0000 (0:00:00.037)       0:01:01.298 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_actual_size": {
        "bytes": 4294967296,
        "changed": false,
        "failed": false,
        "lvm": "4g",
        "parted": "4GiB",
        "size": "4 GiB"
    }
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47
Thursday 21 July 2022  18:53:29 +0000 (0:00:00.038)       0:01:01.337 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50
Thursday 21 July 2022  18:53:30 +0000 (0:00:00.046)       0:01:01.384 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6
Thursday 21 July 2022  18:53:30 +0000 (0:00:00.062)       0:01:01.447 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.027257",
    "end": "2022-07-21 14:53:30.427078",
    "rc": 0,
    "start": "2022-07-21 14:53:30.399821"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14
Thursday 21 July 2022  18:53:30 +0000 (0:00:00.352)       0:01:01.799 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [check segment type] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17
Thursday 21 July 2022  18:53:30 +0000 (0:00:00.052)       0:01:01.852 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22
Thursday 21 July 2022  18:53:30 +0000 (0:00:00.056)       0:01:01.908 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested cache size] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26
Thursday 21 July 2022  18:53:30 +0000 (0:00:00.038)       0:01:01.947 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32
Thursday 21 July 2022  18:53:30 +0000 (0:00:00.039)       0:01:01.987 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36
Thursday 21 July 2022  18:53:30 +0000 (0:00:00.039)       0:01:02.026 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16
Thursday 21 July 2022  18:53:30 +0000 (0:00:00.038)       0:01:02.065 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43
Thursday 21 July 2022  18:53:30 +0000 (0:00:00.036)       0:01:02.102 ********* 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53
Thursday 21 July 2022  18:53:30 +0000 (0:00:00.023)       0:01:02.125 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Resizing with one large value which large than disk's size] **************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:120
Thursday 21 July 2022  18:53:30 +0000 (0:00:00.033)       0:01:02.158 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:53:30 +0000 (0:00:00.040)       0:01:02.198 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:53:30 +0000 (0:00:00.036)       0:01:02.234 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:53:31 +0000 (0:00:00.483)       0:01:02.718 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:53:31 +0000 (0:00:00.061)       0:01:02.779 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:53:31 +0000 (0:00:00.078)       0:01:02.858 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:53:31 +0000 (0:00:00.118)       0:01:02.977 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:53:31 +0000 (0:00:00.059)       0:01:03.036 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:53:31 +0000 (0:00:00.023)       0:01:03.059 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:53:31 +0000 (0:00:00.028)       0:01:03.088 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "sdb"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "12884901888.0"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:53:31 +0000 (0:00:00.052)       0:01:03.140 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:53:31 +0000 (0:00:00.038)       0:01:03.179 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:53:31 +0000 (0:00:00.028)       0:01:03.207 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:53:31 +0000 (0:00:00.027)       0:01:03.235 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:53:31 +0000 (0:00:00.027)       0:01:03.262 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:53:31 +0000 (0:00:00.026)       0:01:03.289 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:53:31 +0000 (0:00:00.059)       0:01:03.349 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:53:32 +0000 (0:00:00.022)       0:01:03.371 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

volume 'test1' cannot be resized to '12 GiB'

TASK [fedora.linux_system_roles.storage : failed message] **********************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:86
Thursday 21 July 2022  18:53:33 +0000 (0:00:01.221)       0:01:04.592 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "changed": false
}

MSG:

{'crypts': [], 'mounts': [], 'leaves': [], 'changed': False, 'actions': [], 'failed': True, 'volumes': [], 'invocation': {'module_args': {'packages_only': False, 'disklabel_type': None, 'diskvolume_mkfs_option_map': {'ext4': '-F', 'ext3': '-F', 'ext2': '-F'}, 'safe_mode': False, 'pools': [{'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', 'size': '12884901888.0', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}], 'volumes': [], 'pool_defaults': {'encryption_password': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_cipher': None, 'disks': [], 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_device_count': None, 'state': 'present', 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'raid_level': None, 'raid_spare_count': None}, 'volume_defaults': {'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'xfs', 'mount_options': 'defaults', 'size': 0, 'mount_point': '', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'fs_overwrite_existing': True, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'mount_passno': 0, 'raid_spare_count': None, 'cache_mode': None, 'deduplication': None, 'cached': False, 'type': 'lvm', 'disks': [], 'thin_pool_size': None, 'thin': None, 'mount_check': 0, 'cache_size': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, 'use_partitions': None}}, 'pools': [], 'packages': [], 'msg': "volume 'test1' cannot be resized to '12 GiB'", '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:53:33 +0000 (0:00:00.042)       0:01:04.635 ********* 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:139
Thursday 21 July 2022  18:53:33 +0000 (0:00:00.022)       0:01:04.657 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the output when resizing with large size] *************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:145
Thursday 21 July 2022  18:53:33 +0000 (0:00:00.040)       0:01:04.698 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Remove the volume group created above] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:150
Thursday 21 July 2022  18:53:33 +0000 (0:00:00.049)       0:01:04.747 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:53:33 +0000 (0:00:00.035)       0:01:04.783 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:53:33 +0000 (0:00:00.034)       0:01:04.818 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:53:33 +0000 (0:00:00.427)       0:01:05.245 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:53:33 +0000 (0:00:00.064)       0:01:05.309 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:53:34 +0000 (0:00:00.070)       0:01:05.380 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:53:34 +0000 (0:00:00.073)       0:01:05.453 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:53:34 +0000 (0:00:00.058)       0:01:05.511 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:53:34 +0000 (0:00:00.022)       0:01:05.534 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:53:34 +0000 (0:00:00.075)       0:01:05.609 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "sdb"
            ],
            "name": "foo",
            "state": "absent",
            "type": "lvm"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:53:34 +0000 (0:00:00.039)       0:01:05.649 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:53:34 +0000 (0:00:00.040)       0:01:05.689 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:53:34 +0000 (0:00:00.032)       0:01:05.722 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:53:34 +0000 (0:00:00.030)       0:01:05.752 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:53:34 +0000 (0:00:00.029)       0:01:05.782 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:53:34 +0000 (0:00:00.030)       0:01:05.812 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:53:34 +0000 (0:00:00.060)       0:01:05.873 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:53:34 +0000 (0:00:00.021)       0:01:05.895 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "destroy device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/sdb",
            "fs_type": "lvmpv"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [
        {
            "fstype": "ext4",
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "absent"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "absent",
            "type": "lvm",
            "volumes": []
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78
Thursday 21 July 2022  18:53:36 +0000 (0:00:01.613)       0:01:07.508 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:53:36 +0000 (0:00:00.036)       0:01:07.545 ********* 

TASK [fedora.linux_system_roles.storage : show blivet_output] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96
Thursday 21 July 2022  18:53:36 +0000 (0:00:00.022)       0:01:07.567 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "destroy device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/sdb",
                "fs_type": "lvmpv"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [
            {
                "fstype": "ext4",
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "absent"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101
Thursday 21 July 2022  18:53:36 +0000 (0:00:00.039)       0:01:07.607 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105
Thursday 21 July 2022  18:53:36 +0000 (0:00:00.037)       0:01:07.644 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : remove obsolete mounts] **************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121
Thursday 21 July 2022  18:53:36 +0000 (0:00:00.037)       0:01:07.681 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'ext4'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "fstype": "ext4",
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Thursday 21 July 2022  18:53:36 +0000 (0:00:00.346)       0:01:08.028 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : set up new/current mounts] ***********
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137
Thursday 21 July 2022  18:53:37 +0000 (0:00:00.495)       0:01:08.524 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Thursday 21 July 2022  18:53:37 +0000 (0:00:00.041)       0:01:08.565 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156
Thursday 21 July 2022  18:53:37 +0000 (0:00:00.464)       0:01:09.030 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429559.3586438,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161
Thursday 21 July 2022  18:53:37 +0000 (0:00:00.319)       0:01:09.350 ********* 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183
Thursday 21 July 2022  18:53:38 +0000 (0:00:00.023)       0:01:09.373 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [Create one partition on one disk] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:160
Thursday 21 July 2022  18:53:38 +0000 (0:00:00.865)       0:01:10.239 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:53:38 +0000 (0:00:00.044)       0:01:10.283 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:53:39 +0000 (0:00:00.073)       0:01:10.357 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:53:39 +0000 (0:00:00.419)       0:01:10.777 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:53:39 +0000 (0:00:00.059)       0:01:10.837 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:53:39 +0000 (0:00:00.038)       0:01:10.875 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:53:39 +0000 (0:00:00.038)       0:01:10.913 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:53:39 +0000 (0:00:00.058)       0:01:10.971 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:53:39 +0000 (0:00:00.026)       0:01:10.998 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:53:39 +0000 (0:00:00.030)       0:01:11.029 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "sdb"
            ],
            "name": "foo",
            "type": "partition",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "type": "partition"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:53:39 +0000 (0:00:00.040)       0:01:11.070 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:53:39 +0000 (0:00:00.037)       0:01:11.108 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:53:39 +0000 (0:00:00.029)       0:01:11.137 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:53:39 +0000 (0:00:00.027)       0:01:11.164 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:53:39 +0000 (0:00:00.031)       0:01:11.196 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:53:39 +0000 (0:00:00.026)       0:01:11.222 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:53:39 +0000 (0:00:00.055)       0:01:11.277 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:53:39 +0000 (0:00:00.021)       0:01:11.299 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "create format",
            "device": "/dev/sdb",
            "fs_type": "disklabel"
        },
        {
            "action": "create device",
            "device": "/dev/sdb1",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/sdb1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd",
        "/dev/sdb1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext4",
            "opts": "defaults",
            "passno": 0,
            "path": "/opt/test1",
            "src": "UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "partition",
            "volumes": [
                {
                    "_device": "/dev/sdb1",
                    "_kernel_device": "/dev/sdb1",
                    "_mount_id": "UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f",
                    "_raw_device": "/dev/sdb1",
                    "_raw_kernel_device": "/dev/sdb1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": 0,
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "partition",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78
Thursday 21 July 2022  18:53:41 +0000 (0:00:01.598)       0:01:12.897 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:53:41 +0000 (0:00:00.037)       0:01:12.935 ********* 

TASK [fedora.linux_system_roles.storage : show blivet_output] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96
Thursday 21 July 2022  18:53:41 +0000 (0:00:00.021)       0:01:12.956 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "create format",
                "device": "/dev/sdb",
                "fs_type": "disklabel"
            },
            {
                "action": "create device",
                "device": "/dev/sdb1",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/sdb1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd",
            "/dev/sdb1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext4",
                "opts": "defaults",
                "passno": 0,
                "path": "/opt/test1",
                "src": "UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "partition",
                "volumes": [
                    {
                        "_device": "/dev/sdb1",
                        "_kernel_device": "/dev/sdb1",
                        "_mount_id": "UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f",
                        "_raw_device": "/dev/sdb1",
                        "_raw_kernel_device": "/dev/sdb1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": 0,
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "partition",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101
Thursday 21 July 2022  18:53:41 +0000 (0:00:00.039)       0:01:12.996 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "partition",
                "volumes": [
                    {
                        "_device": "/dev/sdb1",
                        "_kernel_device": "/dev/sdb1",
                        "_mount_id": "UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f",
                        "_raw_device": "/dev/sdb1",
                        "_raw_kernel_device": "/dev/sdb1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": 0,
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "partition",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105
Thursday 21 July 2022  18:53:41 +0000 (0:00:00.037)       0:01:13.033 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : remove obsolete mounts] **************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121
Thursday 21 July 2022  18:53:41 +0000 (0:00:00.040)       0:01:13.074 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Thursday 21 July 2022  18:53:41 +0000 (0:00:00.036)       0:01:13.111 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : set up new/current mounts] ***********
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137
Thursday 21 July 2022  18:53:42 +0000 (0:00:00.482)       0:01:13.593 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': 'UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f', 'dump': 0, 'passno': 0, 'fstype': 'ext4', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "opts": "defaults",
        "passno": 0,
        "path": "/opt/test1",
        "src": "UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f"
}

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Thursday 21 July 2022  18:53:42 +0000 (0:00:00.357)       0:01:13.950 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156
Thursday 21 July 2022  18:53:43 +0000 (0:00:00.452)       0:01:14.403 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429559.3586438,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161
Thursday 21 July 2022  18:53:43 +0000 (0:00:00.350)       0:01:14.754 ********* 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183
Thursday 21 July 2022  18:53:43 +0000 (0:00:00.021)       0:01:14.776 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:174
Thursday 21 July 2022  18:53:44 +0000 (0:00:00.908)       0:01:15.684 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1
Thursday 21 July 2022  18:53:44 +0000 (0:00:00.038)       0:01:15.722 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "partition",
            "volumes": [
                {
                    "_device": "/dev/sdb1",
                    "_kernel_device": "/dev/sdb1",
                    "_mount_id": "UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f",
                    "_raw_device": "/dev/sdb1",
                    "_raw_kernel_device": "/dev/sdb1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": 0,
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "partition",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6
Thursday 21 July 2022  18:53:44 +0000 (0:00:00.053)       0:01:15.776 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14
Thursday 21 July 2022  18:53:44 +0000 (0:00:00.037)       0:01:15.814 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb1": {
            "fstype": "ext4",
            "label": "",
            "name": "/dev/sdb1",
            "size": "10G",
            "type": "partition",
            "uuid": "a40343b2-9b62-4eb6-87cc-55d04063de3f"
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-18-52-16-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19
Thursday 21 July 2022  18:53:44 +0000 (0:00:00.318)       0:01:16.133 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003530",
    "end": "2022-07-21 14:53:45.074039",
    "rc": 0,
    "start": "2022-07-21 14:53:45.070509"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0
UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f /opt/test1 ext4 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.312)       0:01:16.445 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003412",
    "end": "2022-07-21 14:53:45.380754",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 14:53:45.377342"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.305)       0:01:16.751 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/sdb1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/sdb1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/sdb1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/sdb1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'partition', 'encryption_cipher': None, 'raid_spare_count': None})

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:5
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.061)       0:01:16.812 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:18
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.033)       0:01:16.845 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members)
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes)

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:1
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.047)       0:01:16.893 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:6
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.025)       0:01:16.919 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:15
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.022)       0:01:16.941 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:19
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.024)       0:01:16.966 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify PV count] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:23
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.022)       0:01:16.988 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:29
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.022)       0:01:17.011 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:33
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.022)       0:01:17.033 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:37
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.025)       0:01:17.058 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:41
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.022)       0:01:17.081 ********* 

TASK [Check MD RAID] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:50
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.022)       0:01:17.103 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml for /cache/rhel-7.qcow2

TASK [get information about RAID] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:6
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.042)       0:01:17.145 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:12
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.026)       0:01:17.171 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:16
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.025)       0:01:17.197 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:20
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.025)       0:01:17.222 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:24
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.026)       0:01:17.249 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:30
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.025)       0:01:17.274 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:36
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.024)       0:01:17.299 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:44
Thursday 21 July 2022  18:53:45 +0000 (0:00:00.024)       0:01:17.323 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:53
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.107)       0:01:17.431 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml:1
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.045)       0:01:17.477 ********* 
skipping: [/cache/rhel-7.qcow2] => (item={'_raw_device': '/dev/sdb1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/sdb1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/sdb1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/sdb1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})  => {
    "ansible_loop_var": "storage_test_lvmraid_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_lvmraid_volume": {
        "_device": "/dev/sdb1",
        "_kernel_device": "/dev/sdb1",
        "_mount_id": "UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f",
        "_raw_device": "/dev/sdb1",
        "_raw_kernel_device": "/dev/sdb1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [],
        "encryption": false,
        "encryption_cipher": null,
        "encryption_key": null,
        "encryption_key_size": null,
        "encryption_luks_version": null,
        "encryption_password": null,
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "ext4",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "size": 0,
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:56
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.031)       0:01:17.508 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml for /cache/rhel-7.qcow2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml:1
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.045)       0:01:17.554 ********* 
skipping: [/cache/rhel-7.qcow2] => (item={'_raw_device': '/dev/sdb1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/sdb1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/sdb1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/sdb1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})  => {
    "ansible_loop_var": "storage_test_thin_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_thin_volume": {
        "_device": "/dev/sdb1",
        "_kernel_device": "/dev/sdb1",
        "_mount_id": "UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f",
        "_raw_device": "/dev/sdb1",
        "_raw_kernel_device": "/dev/sdb1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [],
        "encryption": false,
        "encryption_cipher": null,
        "encryption_key": null,
        "encryption_key_size": null,
        "encryption_luks_version": null,
        "encryption_password": null,
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "ext4",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "size": 0,
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [Check member encryption] *************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:59
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.030)       0:01:17.584 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:4
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.045)       0:01:17.629 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:8
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.049)       0:01:17.679 ********* 

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:15
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.029)       0:01:17.709 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:22
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.028)       0:01:17.738 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:62
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.036)       0:01:17.775 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml:1
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.048)       0:01:17.823 ********* 
skipping: [/cache/rhel-7.qcow2] => (item={'_raw_device': '/dev/sdb1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/sdb1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/sdb1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/sdb1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})  => {
    "ansible_loop_var": "storage_test_vdo_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_vdo_volume": {
        "_device": "/dev/sdb1",
        "_kernel_device": "/dev/sdb1",
        "_mount_id": "UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f",
        "_raw_device": "/dev/sdb1",
        "_raw_kernel_device": "/dev/sdb1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [],
        "encryption": false,
        "encryption_cipher": null,
        "encryption_key": null,
        "encryption_key_size": null,
        "encryption_luks_version": null,
        "encryption_password": null,
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "ext4",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "size": 0,
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:65
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.031)       0:01:17.855 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [verify the volumes] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml:3
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.040)       0:01:17.896 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/sdb1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/sdb1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/sdb1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/sdb1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.041)       0:01:17.937 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.053)       0:01:17.990 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache)

TASK [Get expected mount device based on device type] **************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.080)       0:01:18.071 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/sdb1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.042)       0:01:18.113 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": [
            {
                "block_available": 2402892,
                "block_size": 4096,
                "block_total": 2547269,
                "block_used": 144377,
                "device": "/dev/sdb1",
                "fstype": "ext4",
                "inode_available": 655349,
                "inode_total": 655360,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 9842245632,
                "size_total": 10433613824,
                "uuid": "a40343b2-9b62-4eb6-87cc-55d04063de3f"
            }
        ],
        "storage_test_mount_expected_match_count": "1",
        "storage_test_mount_point_matches": [
            {
                "block_available": 2402892,
                "block_size": 4096,
                "block_total": 2547269,
                "block_used": 144377,
                "device": "/dev/sdb1",
                "fstype": "ext4",
                "inode_available": 655349,
                "inode_total": 655360,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 9842245632,
                "size_total": 10433613824,
                "uuid": "a40343b2-9b62-4eb6-87cc-55d04063de3f"
            }
        ],
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.057)       0:01:18.171 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the current mount state by mount point] ***************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.052)       0:01:18.223 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the mount fs type] ************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.048)       0:01:18.272 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [command] *****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.051)       0:01:18.323 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50
Thursday 21 July 2022  18:53:46 +0000 (0:00:00.028)       0:01:18.352 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55
Thursday 21 July 2022  18:53:47 +0000 (0:00:00.026)       0:01:18.378 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65
Thursday 21 July 2022  18:53:47 +0000 (0:00:00.024)       0:01:18.402 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": null,
        "storage_test_mount_expected_match_count": null,
        "storage_test_mount_point_matches": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2
Thursday 21 July 2022  18:53:47 +0000 (0:00:00.037)       0:01:18.440 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12
Thursday 21 July 2022  18:53:47 +0000 (0:00:00.059)       0:01:18.500 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19
Thursday 21 July 2022  18:53:47 +0000 (0:00:00.045)       0:01:18.546 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25
Thursday 21 July 2022  18:53:47 +0000 (0:00:00.047)       0:01:18.593 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up variables] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34
Thursday 21 July 2022  18:53:47 +0000 (0:00:00.045)       0:01:18.638 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4
Thursday 21 July 2022  18:53:47 +0000 (0:00:00.034)       0:01:18.673 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10
Thursday 21 July 2022  18:53:47 +0000 (0:00:00.079)       0:01:18.753 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4
Thursday 21 July 2022  18:53:47 +0000 (0:00:00.081)       0:01:18.834 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429621.4866438,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1658429621.4866438,
        "dev": 5,
        "device_type": 2065,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 48082,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/blockdevice",
        "mode": "0660",
        "mtime": 1658429621.4866438,
        "nlink": 1,
        "path": "/dev/sdb1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10
Thursday 21 July 2022  18:53:47 +0000 (0:00:00.343)       0:01:19.178 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15
Thursday 21 July 2022  18:53:47 +0000 (0:00:00.036)       0:01:19.215 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [(1/2) Process volume type (set initial value)] ***************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21
Thursday 21 July 2022  18:53:47 +0000 (0:00:00.040)       0:01:19.255 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "st_volume_type": "partition"
    },
    "changed": false
}

TASK [(2/2) Process volume type (get RAID value)] ******************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25
Thursday 21 July 2022  18:53:47 +0000 (0:00:00.036)       0:01:19.292 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30
Thursday 21 July 2022  18:53:47 +0000 (0:00:00.023)       0:01:19.316 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3
Thursday 21 July 2022  18:53:47 +0000 (0:00:00.035)       0:01:19.352 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10
Thursday 21 July 2022  18:53:48 +0000 (0:00:00.021)       0:01:19.374 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "rc": 0,
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15
Thursday 21 July 2022  18:53:48 +0000 (0:00:00.542)       0:01:19.917 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21
Thursday 21 July 2022  18:53:48 +0000 (0:00:00.025)       0:01:19.943 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27
Thursday 21 July 2022  18:53:48 +0000 (0:00:00.025)       0:01:19.968 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33
Thursday 21 July 2022  18:53:48 +0000 (0:00:00.052)       0:01:20.021 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39
Thursday 21 July 2022  18:53:48 +0000 (0:00:00.025)       0:01:20.046 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44
Thursday 21 July 2022  18:53:48 +0000 (0:00:00.025)       0:01:20.072 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50
Thursday 21 July 2022  18:53:48 +0000 (0:00:00.025)       0:01:20.097 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56
Thursday 21 July 2022  18:53:48 +0000 (0:00:00.025)       0:01:20.123 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62
Thursday 21 July 2022  18:53:48 +0000 (0:00:00.024)       0:01:20.148 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67
Thursday 21 July 2022  18:53:48 +0000 (0:00:00.053)       0:01:20.202 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72
Thursday 21 July 2022  18:53:48 +0000 (0:00:00.054)       0:01:20.256 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78
Thursday 21 July 2022  18:53:48 +0000 (0:00:00.037)       0:01:20.293 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84
Thursday 21 July 2022  18:53:48 +0000 (0:00:00.037)       0:01:20.331 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.038)       0:01:20.369 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [get information about RAID] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.039)       0:01:20.409 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.037)       0:01:20.447 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.035)       0:01:20.482 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.035)       0:01:20.518 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.039)       0:01:20.558 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.036)       0:01:20.594 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.036)       0:01:20.631 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the actual size of the volume] *************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.035)       0:01:20.666 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested size of the volume] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.025)       0:01:20.692 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.035)       0:01:20.727 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.035)       0:01:20.763 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.036)       0:01:20.799 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.040)       0:01:20.839 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.038)       0:01:20.877 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.037)       0:01:20.915 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.036)       0:01:20.951 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.040)       0:01:20.992 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.036)       0:01:21.028 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.071)       0:01:21.100 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.025)       0:01:21.126 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.022)       0:01:21.149 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check segment type] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.022)       0:01:21.171 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.023)       0:01:21.194 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested cache size] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.027)       0:01:21.221 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.024)       0:01:21.246 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.024)       0:01:21.271 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16
Thursday 21 July 2022  18:53:49 +0000 (0:00:00.024)       0:01:21.295 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43
Thursday 21 July 2022  18:53:50 +0000 (0:00:00.106)       0:01:21.402 ********* 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53
Thursday 21 July 2022  18:53:50 +0000 (0:00:00.022)       0:01:21.424 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Test setting up disk volume will remove the partition create above] ******
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:176
Thursday 21 July 2022  18:53:50 +0000 (0:00:00.034)       0:01:21.459 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:53:50 +0000 (0:00:00.046)       0:01:21.506 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:53:50 +0000 (0:00:00.034)       0:01:21.540 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:53:50 +0000 (0:00:00.425)       0:01:21.966 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:53:50 +0000 (0:00:00.064)       0:01:22.030 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:53:50 +0000 (0:00:00.037)       0:01:22.068 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:53:50 +0000 (0:00:00.036)       0:01:22.105 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:53:50 +0000 (0:00:00.055)       0:01:22.160 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:53:50 +0000 (0:00:00.021)       0:01:22.182 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:53:50 +0000 (0:00:00.027)       0:01:22.209 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:53:50 +0000 (0:00:00.037)       0:01:22.246 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": [
        {
            "disks": [
                "sdb"
            ],
            "fs_create_options": "-F",
            "fs_type": "ext4",
            "mount_options": "rw,noatime,defaults",
            "mount_point": "/opt/test1",
            "name": "foo",
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:53:50 +0000 (0:00:00.040)       0:01:22.287 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:53:50 +0000 (0:00:00.029)       0:01:22.317 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:53:50 +0000 (0:00:00.029)       0:01:22.347 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:53:51 +0000 (0:00:00.029)       0:01:22.376 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:53:51 +0000 (0:00:00.029)       0:01:22.406 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:53:51 +0000 (0:00:00.062)       0:01:22.468 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:53:51 +0000 (0:00:00.023)       0:01:22.492 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/sdb1",
            "fs_type": "ext4"
        },
        {
            "action": "destroy device",
            "device": "/dev/sdb1",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/sdb",
            "fs_type": "disklabel"
        },
        {
            "action": "create format",
            "device": "/dev/sdb",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [
        {
            "fstype": "ext4",
            "path": "/opt/test1",
            "src": "UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f",
            "state": "absent"
        },
        {
            "dump": 0,
            "fstype": "ext4",
            "opts": "rw,noatime,defaults",
            "passno": 0,
            "path": "/opt/test1",
            "src": "UUID=91c9f21a-206d-4051-86cf-2b2b90e96331",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs"
    ],
    "pools": [],
    "volumes": [
        {
            "_device": "/dev/sdb",
            "_kernel_device": "/dev/sdb",
            "_mount_id": "UUID=91c9f21a-206d-4051-86cf-2b2b90e96331",
            "_raw_device": "/dev/sdb",
            "_raw_kernel_device": "/dev/sdb",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "fs_create_options": "-F",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "ext4",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_options": "rw,noatime,defaults",
            "mount_passno": 0,
            "mount_point": "/opt/test1",
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "size": 10737418240,
            "state": "present",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78
Thursday 21 July 2022  18:53:52 +0000 (0:00:01.663)       0:01:24.156 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:53:52 +0000 (0:00:00.040)       0:01:24.196 ********* 

TASK [fedora.linux_system_roles.storage : show blivet_output] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96
Thursday 21 July 2022  18:53:52 +0000 (0:00:00.022)       0:01:24.219 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/sdb1",
                "fs_type": "ext4"
            },
            {
                "action": "destroy device",
                "device": "/dev/sdb1",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/sdb",
                "fs_type": "disklabel"
            },
            {
                "action": "create format",
                "device": "/dev/sdb",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [
            {
                "fstype": "ext4",
                "path": "/opt/test1",
                "src": "UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f",
                "state": "absent"
            },
            {
                "dump": 0,
                "fstype": "ext4",
                "opts": "rw,noatime,defaults",
                "passno": 0,
                "path": "/opt/test1",
                "src": "UUID=91c9f21a-206d-4051-86cf-2b2b90e96331",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs"
        ],
        "pools": [],
        "volumes": [
            {
                "_device": "/dev/sdb",
                "_kernel_device": "/dev/sdb",
                "_mount_id": "UUID=91c9f21a-206d-4051-86cf-2b2b90e96331",
                "_raw_device": "/dev/sdb",
                "_raw_kernel_device": "/dev/sdb",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "fs_create_options": "-F",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "ext4",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_options": "rw,noatime,defaults",
                "mount_passno": 0,
                "mount_point": "/opt/test1",
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "size": 10737418240,
                "state": "present",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    }
}

TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101
Thursday 21 July 2022  18:53:52 +0000 (0:00:00.050)       0:01:24.269 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105
Thursday 21 July 2022  18:53:53 +0000 (0:00:00.084)       0:01:24.354 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": [
            {
                "_device": "/dev/sdb",
                "_kernel_device": "/dev/sdb",
                "_mount_id": "UUID=91c9f21a-206d-4051-86cf-2b2b90e96331",
                "_raw_device": "/dev/sdb",
                "_raw_kernel_device": "/dev/sdb",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "fs_create_options": "-F",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "ext4",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_options": "rw,noatime,defaults",
                "mount_passno": 0,
                "mount_point": "/opt/test1",
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "size": 10737418240,
                "state": "present",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : remove obsolete mounts] **************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121
Thursday 21 July 2022  18:53:53 +0000 (0:00:00.037)       0:01:24.391 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': 'UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'ext4'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "fstype": "ext4",
        "path": "/opt/test1",
        "src": "UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "UUID=a40343b2-9b62-4eb6-87cc-55d04063de3f"
}

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Thursday 21 July 2022  18:53:53 +0000 (0:00:00.400)       0:01:24.791 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : set up new/current mounts] ***********
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137
Thursday 21 July 2022  18:53:53 +0000 (0:00:00.516)       0:01:25.308 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': 'UUID=91c9f21a-206d-4051-86cf-2b2b90e96331', 'dump': 0, 'passno': 0, 'fstype': 'ext4', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'rw,noatime,defaults'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "opts": "rw,noatime,defaults",
        "passno": 0,
        "path": "/opt/test1",
        "src": "UUID=91c9f21a-206d-4051-86cf-2b2b90e96331",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "rw,noatime,defaults",
    "passno": "0",
    "src": "UUID=91c9f21a-206d-4051-86cf-2b2b90e96331"
}

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Thursday 21 July 2022  18:53:54 +0000 (0:00:00.385)       0:01:25.693 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156
Thursday 21 July 2022  18:53:54 +0000 (0:00:00.454)       0:01:26.148 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429559.3586438,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161
Thursday 21 July 2022  18:53:55 +0000 (0:00:00.327)       0:01:26.475 ********* 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183
Thursday 21 July 2022  18:53:55 +0000 (0:00:00.022)       0:01:26.497 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:189
Thursday 21 July 2022  18:53:55 +0000 (0:00:00.852)       0:01:27.349 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1
Thursday 21 July 2022  18:53:56 +0000 (0:00:00.044)       0:01:27.394 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Print out volume information] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6
Thursday 21 July 2022  18:53:56 +0000 (0:00:00.040)       0:01:27.434 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_volumes_list": [
        {
            "_device": "/dev/sdb",
            "_kernel_device": "/dev/sdb",
            "_mount_id": "UUID=91c9f21a-206d-4051-86cf-2b2b90e96331",
            "_raw_device": "/dev/sdb",
            "_raw_kernel_device": "/dev/sdb",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "fs_create_options": "-F",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "ext4",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_options": "rw,noatime,defaults",
            "mount_passno": 0,
            "mount_point": "/opt/test1",
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "size": 10737418240,
            "state": "present",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14
Thursday 21 July 2022  18:53:56 +0000 (0:00:00.059)       0:01:27.494 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "ext4",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": "91c9f21a-206d-4051-86cf-2b2b90e96331"
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-18-52-16-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19
Thursday 21 July 2022  18:53:56 +0000 (0:00:00.321)       0:01:27.815 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003582",
    "end": "2022-07-21 14:53:56.758042",
    "rc": 0,
    "start": "2022-07-21 14:53:56.754460"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0
UUID=91c9f21a-206d-4051-86cf-2b2b90e96331 /opt/test1 ext4 rw,noatime,defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24
Thursday 21 July 2022  18:53:56 +0000 (0:00:00.312)       0:01:28.127 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003619",
    "end": "2022-07-21 14:53:57.079845",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 14:53:57.076226"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33
Thursday 21 July 2022  18:53:57 +0000 (0:00:00.325)       0:01:28.452 ********* 

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43
Thursday 21 July 2022  18:53:57 +0000 (0:00:00.021)       0:01:28.474 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/sdb', 'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'ext4', 'mount_options': 'rw,noatime,defaults', '_device': '/dev/sdb', 'size': 10737418240, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/sdb', 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'type': 'disk', 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=91c9f21a-206d-4051-86cf-2b2b90e96331', 'raid_spare_count': None, 'name': 'foo', '_raw_kernel_device': '/dev/sdb', 'cache_mode': None, 'cache_devices': [], 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': None, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'thin_pool_size': None, 'fs_create_options': '-F'})

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2
Thursday 21 July 2022  18:53:57 +0000 (0:00:00.058)       0:01:28.533 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10
Thursday 21 July 2022  18:53:57 +0000 (0:00:00.048)       0:01:28.582 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache)

TASK [Get expected mount device based on device type] **************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6
Thursday 21 July 2022  18:53:57 +0000 (0:00:00.089)       0:01:28.672 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/sdb"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10
Thursday 21 July 2022  18:53:57 +0000 (0:00:00.090)       0:01:28.762 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": [
            {
                "block_available": 2403135,
                "block_size": 4096,
                "block_total": 2547525,
                "block_used": 144390,
                "device": "/dev/sdb",
                "fstype": "ext4",
                "inode_available": 655349,
                "inode_total": 655360,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,noatime,data=ordered",
                "size_available": 9843240960,
                "size_total": 10434662400,
                "uuid": "91c9f21a-206d-4051-86cf-2b2b90e96331"
            }
        ],
        "storage_test_mount_expected_match_count": "1",
        "storage_test_mount_point_matches": [
            {
                "block_available": 2403135,
                "block_size": 4096,
                "block_total": 2547525,
                "block_used": 144390,
                "device": "/dev/sdb",
                "fstype": "ext4",
                "inode_available": 655349,
                "inode_total": 655360,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,noatime,data=ordered",
                "size_available": 9843240960,
                "size_total": 10434662400,
                "uuid": "91c9f21a-206d-4051-86cf-2b2b90e96331"
            }
        ],
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20
Thursday 21 July 2022  18:53:57 +0000 (0:00:00.102)       0:01:28.865 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the current mount state by mount point] ***************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29
Thursday 21 July 2022  18:53:57 +0000 (0:00:00.087)       0:01:28.953 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the mount fs type] ************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37
Thursday 21 July 2022  18:53:57 +0000 (0:00:00.084)       0:01:29.037 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [command] *****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46
Thursday 21 July 2022  18:53:57 +0000 (0:00:00.055)       0:01:29.092 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50
Thursday 21 July 2022  18:53:57 +0000 (0:00:00.022)       0:01:29.115 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55
Thursday 21 July 2022  18:53:57 +0000 (0:00:00.036)       0:01:29.152 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65
Thursday 21 July 2022  18:53:57 +0000 (0:00:00.025)       0:01:29.177 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": null,
        "storage_test_mount_expected_match_count": null,
        "storage_test_mount_point_matches": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2
Thursday 21 July 2022  18:53:57 +0000 (0:00:00.033)       0:01:29.211 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "UUID=91c9f21a-206d-4051-86cf-2b2b90e96331 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 rw,noatime,defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12
Thursday 21 July 2022  18:53:57 +0000 (0:00:00.064)       0:01:29.276 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19
Thursday 21 July 2022  18:53:57 +0000 (0:00:00.056)       0:01:29.332 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25
Thursday 21 July 2022  18:53:58 +0000 (0:00:00.046)       0:01:29.379 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34
Thursday 21 July 2022  18:53:58 +0000 (0:00:00.051)       0:01:29.431 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4
Thursday 21 July 2022  18:53:58 +0000 (0:00:00.034)       0:01:29.465 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10
Thursday 21 July 2022  18:53:58 +0000 (0:00:00.035)       0:01:29.501 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4
Thursday 21 July 2022  18:53:58 +0000 (0:00:00.049)       0:01:29.550 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429632.7356436,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1658429632.7356436,
        "dev": 5,
        "device_type": 2064,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 11712,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/blockdevice",
        "mode": "0660",
        "mtime": 1658429632.7356436,
        "nlink": 1,
        "path": "/dev/sdb",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10
Thursday 21 July 2022  18:53:58 +0000 (0:00:00.326)       0:01:29.877 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15
Thursday 21 July 2022  18:53:58 +0000 (0:00:00.041)       0:01:29.918 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [(1/2) Process volume type (set initial value)] ***************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21
Thursday 21 July 2022  18:53:58 +0000 (0:00:00.039)       0:01:29.957 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "st_volume_type": "disk"
    },
    "changed": false
}

TASK [(2/2) Process volume type (get RAID value)] ******************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25
Thursday 21 July 2022  18:53:58 +0000 (0:00:00.033)       0:01:29.991 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30
Thursday 21 July 2022  18:53:58 +0000 (0:00:00.021)       0:01:30.013 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3
Thursday 21 July 2022  18:53:58 +0000 (0:00:00.036)       0:01:30.049 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10
Thursday 21 July 2022  18:53:58 +0000 (0:00:00.025)       0:01:30.074 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "rc": 0,
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.548)       0:01:30.623 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.025)       0:01:30.648 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.025)       0:01:30.674 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.053)       0:01:30.727 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.026)       0:01:30.754 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.024)       0:01:30.779 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.024)       0:01:30.803 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.024)       0:01:30.827 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.025)       0:01:30.853 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.056)       0:01:30.910 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.050)       0:01:30.960 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.039)       0:01:31.000 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.035)       0:01:31.035 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.037)       0:01:31.072 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [get information about RAID] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.036)       0:01:31.108 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.042)       0:01:31.151 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.038)       0:01:31.189 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.037)       0:01:31.227 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.036)       0:01:31.263 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.041)       0:01:31.304 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37
Thursday 21 July 2022  18:53:59 +0000 (0:00:00.037)       0:01:31.342 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the actual size of the volume] *************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.035)       0:01:31.377 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested size of the volume] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.024)       0:01:31.402 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.074)       0:01:31.477 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.036)       0:01:31.513 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.038)       0:01:31.552 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.037)       0:01:31.589 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.036)       0:01:31.626 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.036)       0:01:31.663 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.039)       0:01:31.702 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.042)       0:01:31.745 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.041)       0:01:31.786 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.034)       0:01:31.820 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.022)       0:01:31.842 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.022)       0:01:31.865 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check segment type] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.024)       0:01:31.890 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.022)       0:01:31.912 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested cache size] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.022)       0:01:31.935 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.021)       0:01:31.957 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.026)       0:01:31.984 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.026)       0:01:32.010 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.034)       0:01:32.045 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Remove the disk volume created above] ************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:193
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.033)       0:01:32.078 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.050)       0:01:32.129 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:54:00 +0000 (0:00:00.034)       0:01:32.164 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:54:01 +0000 (0:00:00.425)       0:01:32.589 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:54:01 +0000 (0:00:00.061)       0:01:32.651 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:54:01 +0000 (0:00:00.035)       0:01:32.686 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:54:01 +0000 (0:00:00.035)       0:01:32.721 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:54:01 +0000 (0:00:00.056)       0:01:32.778 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:54:01 +0000 (0:00:00.019)       0:01:32.798 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:54:01 +0000 (0:00:00.026)       0:01:32.824 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:54:01 +0000 (0:00:00.037)       0:01:32.862 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": [
        {
            "disks": [
                "sdb"
            ],
            "name": "foo",
            "state": "absent",
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:54:01 +0000 (0:00:00.042)       0:01:32.904 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:54:01 +0000 (0:00:00.030)       0:01:32.935 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:54:01 +0000 (0:00:00.029)       0:01:32.964 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:54:01 +0000 (0:00:00.030)       0:01:32.994 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:54:01 +0000 (0:00:00.030)       0:01:33.025 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:54:01 +0000 (0:00:00.091)       0:01:33.117 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:54:01 +0000 (0:00:00.020)       0:01:33.137 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/sdb",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [
        {
            "fstype": "ext4",
            "path": "/opt/test1",
            "src": "UUID=91c9f21a-206d-4051-86cf-2b2b90e96331",
            "state": "absent"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [],
    "volumes": [
        {
            "_device": "/dev/sdb",
            "_mount_id": "UUID=91c9f21a-206d-4051-86cf-2b2b90e96331",
            "_raw_device": "/dev/sdb",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "sdb"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "fs_create_options": "",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "ext4",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_options": "defaults",
            "mount_passno": 0,
            "mount_point": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "size": 10737418240,
            "state": "absent",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78
Thursday 21 July 2022  18:54:02 +0000 (0:00:01.072)       0:01:34.209 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:54:02 +0000 (0:00:00.036)       0:01:34.246 ********* 

TASK [fedora.linux_system_roles.storage : show blivet_output] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96
Thursday 21 July 2022  18:54:02 +0000 (0:00:00.020)       0:01:34.267 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/sdb",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [
            {
                "fstype": "ext4",
                "path": "/opt/test1",
                "src": "UUID=91c9f21a-206d-4051-86cf-2b2b90e96331",
                "state": "absent"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [],
        "volumes": [
            {
                "_device": "/dev/sdb",
                "_mount_id": "UUID=91c9f21a-206d-4051-86cf-2b2b90e96331",
                "_raw_device": "/dev/sdb",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "fs_create_options": "",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "ext4",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_options": "defaults",
                "mount_passno": 0,
                "mount_point": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "size": 10737418240,
                "state": "absent",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    }
}

TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101
Thursday 21 July 2022  18:54:02 +0000 (0:00:00.038)       0:01:34.306 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105
Thursday 21 July 2022  18:54:02 +0000 (0:00:00.038)       0:01:34.344 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": [
            {
                "_device": "/dev/sdb",
                "_mount_id": "UUID=91c9f21a-206d-4051-86cf-2b2b90e96331",
                "_raw_device": "/dev/sdb",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "sdb"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "fs_create_options": "",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "ext4",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_options": "defaults",
                "mount_passno": 0,
                "mount_point": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "size": 10737418240,
                "state": "absent",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : remove obsolete mounts] **************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121
Thursday 21 July 2022  18:54:03 +0000 (0:00:00.036)       0:01:34.380 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': 'UUID=91c9f21a-206d-4051-86cf-2b2b90e96331', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'ext4'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "fstype": "ext4",
        "path": "/opt/test1",
        "src": "UUID=91c9f21a-206d-4051-86cf-2b2b90e96331",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "UUID=91c9f21a-206d-4051-86cf-2b2b90e96331"
}

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Thursday 21 July 2022  18:54:03 +0000 (0:00:00.347)       0:01:34.727 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : set up new/current mounts] ***********
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137
Thursday 21 July 2022  18:54:03 +0000 (0:00:00.456)       0:01:35.184 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Thursday 21 July 2022  18:54:03 +0000 (0:00:00.039)       0:01:35.224 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156
Thursday 21 July 2022  18:54:04 +0000 (0:00:00.444)       0:01:35.668 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429559.3586438,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161
Thursday 21 July 2022  18:54:04 +0000 (0:00:00.326)       0:01:35.995 ********* 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183
Thursday 21 July 2022  18:54:04 +0000 (0:00:00.023)       0:01:36.018 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [Try to mount swap filesystem to "/opt/test1"] ****************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:205
Thursday 21 July 2022  18:54:05 +0000 (0:00:00.836)       0:01:36.855 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:54:05 +0000 (0:00:00.041)       0:01:36.897 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:54:05 +0000 (0:00:00.036)       0:01:36.933 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:54:05 +0000 (0:00:00.418)       0:01:37.351 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:54:06 +0000 (0:00:00.064)       0:01:37.415 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:54:06 +0000 (0:00:00.035)       0:01:37.451 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:54:06 +0000 (0:00:00.070)       0:01:37.522 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:54:06 +0000 (0:00:00.055)       0:01:37.578 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:54:06 +0000 (0:00:00.022)       0:01:37.600 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:54:06 +0000 (0:00:00.030)       0:01:37.630 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:54:06 +0000 (0:00:00.037)       0:01:37.668 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": [
        {
            "disks": [
                "sdb"
            ],
            "fs_type": "swap",
            "mount_point": "/opt/test1",
            "name": "test1",
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:54:06 +0000 (0:00:00.074)       0:01:37.743 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:54:06 +0000 (0:00:00.030)       0:01:37.773 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:54:06 +0000 (0:00:00.069)       0:01:37.842 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:54:06 +0000 (0:00:00.031)       0:01:37.874 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:54:06 +0000 (0:00:00.027)       0:01:37.902 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:54:06 +0000 (0:00:00.057)       0:01:37.960 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:54:06 +0000 (0:00:00.024)       0:01:37.985 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

volume 'test1' has a mount point but no mountable file system

TASK [fedora.linux_system_roles.storage : failed message] **********************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:86
Thursday 21 July 2022  18:54:07 +0000 (0:00:00.895)       0:01:38.881 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "changed": false
}

MSG:

{'crypts': [], 'mounts': [], 'leaves': [], 'changed': False, 'actions': [], 'failed': True, 'volumes': [], 'invocation': {'module_args': {'packages_only': False, 'disklabel_type': None, 'diskvolume_mkfs_option_map': {'ext4': '-F', 'ext3': '-F', 'ext2': '-F'}, 'safe_mode': False, 'pools': [], 'volumes': [{'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'swap', 'mount_options': 'defaults', 'size': 10737418240, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'type': 'disk', 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, 'raid_spare_count': None, 'name': 'test1', 'cache_mode': None, 'cache_devices': [], 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': None, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'thin_pool_size': None, 'fs_create_options': ''}], 'pool_defaults': {'encryption_password': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_cipher': None, 'disks': [], 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_device_count': None, 'state': 'present', 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'raid_level': None, 'raid_spare_count': None}, 'volume_defaults': {'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'xfs', 'mount_options': 'defaults', 'size': 0, 'mount_point': '', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'fs_overwrite_existing': True, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'mount_passno': 0, 'raid_spare_count': None, 'cache_mode': None, 'deduplication': None, 'cached': False, 'type': 'lvm', 'disks': [], 'thin_pool_size': None, 'thin': None, 'mount_check': 0, 'cache_size': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, 'use_partitions': None}}, 'pools': [], 'packages': [], 'msg': "volume 'test1' has a mount point but no mountable file system", '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:54:07 +0000 (0:00:00.042)       0:01:38.923 ********* 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:221
Thursday 21 July 2022  18:54:07 +0000 (0:00:00.020)       0:01:38.944 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the output when mount swap filesystem to "/opt/test1"] ************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:227
Thursday 21 July 2022  18:54:07 +0000 (0:00:00.036)       0:01:38.980 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed
META: ran handlers
META: ran handlers

PLAY RECAP *********************************************************************
/cache/rhel-7.qcow2        : ok=483  changed=19   unreachable=0    failed=3    skipped=367  rescued=3    ignored=0   

Thursday 21 July 2022  18:54:07 +0000 (0:00:00.058)       0:01:39.039 ********* 
=============================================================================== 
fedora.linux_system_roles.storage : make sure blivet is available ------- 8.96s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.73s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.66s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.65s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
Ensure cryptsetup is present -------------------------------------------- 1.62s
/tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 -----------
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.61s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.60s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.59s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
set up internal repositories -------------------------------------------- 1.42s
/cache/rhel-7_setup.yml:5 -----------------------------------------------------
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.39s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.30s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.22s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
fedora.linux_system_roles.storage : get service facts ------------------- 1.14s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.07s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
Gathering Facts --------------------------------------------------------- 1.03s
/tmp/tmptomayb7j/tests/storage/tests_misc_scsi_generated.yml:3 ----------------
fedora.linux_system_roles.storage : Update facts ------------------------ 0.91s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 
fedora.linux_system_roles.storage : Update facts ------------------------ 0.90s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 0.90s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
fedora.linux_system_roles.storage : Update facts ------------------------ 0.88s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 
fedora.linux_system_roles.storage : Update facts ------------------------ 0.87s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183